Papers by Jose Parente de Oliveira
In the context of software dependability, the software verification process has an important role... more In the context of software dependability, the software verification process has an important role. Formal verification of programs is an activity that can be inserted in this process to improve software reliability. This paper presents the definition of an approach that employs a formal verification technique based on abstract interpretation. The main goal is to apply this technique as a formal activity in the software verification process to help software engineers identify programs faults. The applicability of the proposed approach is demonstrated by a case study based on embedded aerospace control software. The results obtained from its use show that abstract interpretation can contribute to software dependability.

IEEE Access, 2021
The increase of connected devices and the constantly evolving methods and techniques by attackers... more The increase of connected devices and the constantly evolving methods and techniques by attackers pose a challenge for network intrusion detection systems from conception to operation. As a result, we see a constant adoption of machine learning algorithms for network intrusion detection systems. However, the dataset used by these studies has become obsolete regarding both background and attack traffic. This work describes the AB-TRAP framework that enables the use of updated network traffic and considers operational concerns to enable the complete deployment of the solution. AB-TRAP is a five-step framework consisting of (i) the generation of the attack dataset, (ii) the bonafide dataset, (iii) training of machine learning models, (iv) realization (implementation) of the models, and (v) the performance evaluation of the realized model after deployment. We exercised the AB-TRAP for local (LAN) and global (internet) environments to detect TCP port scanning attacks. The LAN study case presented an f1-score of 0.96, and an area under the ROC curve of 0.99 using a decision tree with minimal CPU and RAM usage on kernel-space. For the internet case with eight machine learning algorithms with an average f1-score of 0.95, an average area under the ROC curve of 0.98, and an average overhead of 1.4% CPU and 3.6% RAM on user-space in a single-board computer. This framework has the following paramount characteristics: it is reproducible, uses the most up-to-date network traffic, attacks, and addresses the concerns to the model's realization and deployment. INDEX TERMS Cybersecurity, datasets, intrusion detection system, machine learning, network security, supervised learning. NOMENCLATURE AUC Area Under the Curve. B5G 5G and beyond. CoAP Constrained Application Protocol. CPS Cyber-Physical System. CSV Comma-separated values. DDoS Distributed Denial of Service. DL Deep Learning. DNP3 Distributed Network Protocol 3. DoS Denial of Service. The associate editor coordinating the review of this manuscript and approving it for publication was Muhammad Asif .
In the context of software dependability, the software verification process has an important role... more In the context of software dependability, the software verification process has an important role. Formal verification of programs is an activity that can be inserted in this process to improve software reliability. This paper presents the definition of an approach that employs a formal verification technique based on abstract interpretation. The main goal is to apply this technique as a formal activity in the software verification process to help software engineers identify programs faults. The applicability of the proposed approach is demonstrated by a case study based on embedded aerospace control software. The results obtained from its use show that abstract interpretation can contribute to software dependability.
2012 IEEE/AIAA 31st Digital Avionics Systems Conference (DASC), 2012

The integration of different datasets in the Linked Data Cloud is a key aspect to the success of ... more The integration of different datasets in the Linked Data Cloud is a key aspect to the success of the Web of Data. To tackle this problem most of existent solutions have been supported by the task of entity resolution. However, many challenges still prevail specially when considering different types, structures and vocabularies used in the Web. Another common problem is that data usually are incomplete, inconsistent and contain outliers. To overcome these limitations, some works have applied machine learning algorithms since they are typically robust to both noise and data inconsistencies and are able to efficiently utilize nondeterministic dependencies in the data. In this paper we propose an approach based in a relational learning algorithm that addresses the problem by statistical approximation method. Modeling the problem as a relational machine learning task allows exploit contextual information that might be too distant in the relational graph. The joint application of relation...
The design and implementation of a hypertext environment for supporting the development of traini... more The design and implementation of a hypertext environment for supporting the development of training courses according to the TRAINAIR Methodology, named TrainTool, is presented in this paper. The purpose of this environment is to improve the process of handling, storing, and retrieving information compiled during a course preparation. To specify such a supporting environment, a model for modeling hypertext applications
International Journal of Computer Applications, 2012
With the documents increasing amount available in local or Web repositories, the comparison metho... more With the documents increasing amount available in local or Web repositories, the comparison methods have to analyze large documents sets with different types and terminologies to obtain a response with minimum documents and with as much useful content to the user. For large documents sets where each document can contain many pages, it is impossible to compute the similarity using the entire document, to require creating solutions to analyze a few meaningful terms, in summary form. This article presents TextSSimily, a method that compares documents semantically considering only short text for comparison (text summary), using semantics to improve the set of responses and summaries to improve time to obtain results for large sets of documents.

IEEE Transactions on Reliability, 2016
With the increasing importance of software in the aerospace field, as evidenced by its growing si... more With the increasing importance of software in the aerospace field, as evidenced by its growing size and complexity, a rigorous and reliable software verification and validation process must be applied to ensure conformance with the strict requirements of this software. Although important, traditional validation activities such as testing and simulation can only provide a partial verification of behavior in critical real-time software systems, and thus, formal verification is an alternative to complement these activities. Two useful formal software verification approaches are deductive verification and abstract interpretation, which analyze programs statically to identify defects. This paper explores abstract interpretation and deductive verification by employing Frama-C's value analysis and Jessie plug-ins to verify embedded aerospace control software. The results indicate that both approaches can be employed in a software verification process to make software more reliable.
Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering, 2012
Problem solving requires a lot of work for selecting and organizing resources and information, as... more Problem solving requires a lot of work for selecting and organizing resources and information, as well as defining the best solution approaches. In the case of ill-structured problems (ISP), due to the possibility of several potential solutions, the task becomes harder. Group work can better support this task but requires cognitive tools in order to allow the registration and the recovery of data about the problem and the exchange of ideas. This paper presents a process to support solving ISP, using ontology and software tools in order to facilitate the group work. Preliminary experiments indicate the feasibility of the process.

J. Univers. Comput. Sci., 2011
Web Services have become a standard for integration of systems in dis- tributed environments. By ... more Web Services have become a standard for integration of systems in dis- tributed environments. By using a set of open interoperability standards, they allow computer-computer interaction, regardless the programming languages and operating systems used. The Semantic Web Services, by its turn, make use of ontologies to de- scribe their functionality in a more structural manner, allowing computers to reason about the information required and provided by them. Such a description also allows dynamic composition of several Web Services, when only one is not able to provide the desired functionality. There are scenarios, however, in which only the functional cor- rectness is not enough to fulfill the user requirements, and a minimum level of quality should be guaranteed by their providers. In this context, this work presents an ap- proach for dynamic Web Service composition that takes into account the composition overall quality. The proposed approach relies on a heuristics to efficiently p...

2018 IEEE First International Conference on Artificial Intelligence and Knowledge Engineering (AIKE), 2018
Knowledge graph completion is still a challenging solution that uses techniques from distinct are... more Knowledge graph completion is still a challenging solution that uses techniques from distinct areas to solve many different tasks. Most recent works, which are based on embedding models, were conceived to improve an existing knowledge graph using the link prediction task. However, even considering the ability of these solutions to solve other tasks, they did not present results for data linking, for example. Furthermore, most of these works focuses only on structural information, i.e., the relations between entities. In this paper, we present an approach for data linking that enrich entity embeddings in a model with their literal information and that do not rely on external information of these entities. The key aspect of this proposal is that we use a blocking scheme to improve the effectiveness of the solution in relation to the use of literals. Thus, in addition to the literals from object elements in a triple, we use other literals from subjects and predicates. By merging entity embeddings with their literal information it is possible to extend many popular embedding models. Preliminary experiments were performed on real-world datasets and our solution showed competitive results to the performance of the task of data linking.
Anais Do Simposio Brasileiro De Informatica Na Educacao, 2012

2016 IEEE 30th International Conference on Advanced Information Networking and Applications (AINA), 2016
The amount and diversity of data in the Semantic Web has grown quite. RDF datasets has proportion... more The amount and diversity of data in the Semantic Web has grown quite. RDF datasets has proportionally more problems than relational datasets due to the way data are published, usually without formal criteria. Entity Resolution is n important issue which is related to a known task of many research communities and it aims at finding all representations that refer to the same entity in different datasets. Yet, it is still an open problem. Blocking methods are used to avoid the quadratic complexity of the brute force approach by clustering entities into blocks and limiting the evaluation of entity specifications to entity pairs within blocks. In the last years only a few blocking methods were conceived to deal with RDF data and novel blocking techniques are required for dealing with noisy and heterogeneous data in the Web of Data. In this paper we present a blocking scheme, CER-Blocking, which is based on an inverted index structure and that uses different data evidences from a triple, aiming to maximize its effectiveness. To overcome the problems of data quality or even the very absence thereof, we use two blocking key definitions. This scheme is part of an ER approach which is based on a relational learning algorithm that addresses the problem by statistical approximation. It was empirically evaluated on real and synthetic datasets which are part of consolidated benchmarks found on the literature.
Anais Do Simposio Brasileiro De Informatica Na Educacao, 2011

2015 11th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS), 2015
The low success rate when retrieving information through web searches could be verified virtually... more The low success rate when retrieving information through web searches could be verified virtually in all areas of knowledge, due to the large amount of information available which raises the selection complexity for relevant articles. A query consists in chosen terms to drive the search for related documents. However, if new terms could be added in order to expand the relevance of the search, then there is what is called query semantic enrichment. This paper presents a semantic enrichment model to improve the quality of results for medical articles queries. This model knows the search context by using a repository of articles which is previously subjected to Latent Semantic Analysis and is supported by the National Cancer Institute ontology and the WordNet lexical database. In this way, new terms which are semantically related to the conducted search context, could be proposed to help raising precision when retrieving relevant articles.
Adaptive Educational Hypermedia Systems (AEHS) have been used to support customized learning. The... more Adaptive Educational Hypermedia Systems (AEHS) have been used to support customized learning. The adaptation mechanisms provided usually try to define the better concept sequence to be presented and to select the materials and activities more appropriate for a given learner. Nevertheless, despite the primary purpose of AEHS in supporting learning, adaptation mechanisms in these systems have no compromise in using an instructional design theory as source of information. This paper addresses a way in which an instructional design or learning theory is used in conjunction with learner's domain knowledge, background, preferences and learning style as source of information for adaptation. To make that viable, a framework for AEHS was defined. The paper describes the framework's features and its potential benefits.

Este artigo aborda algumas iniciativas e tendências das aplicações de Web Semântica no governo el... more Este artigo aborda algumas iniciativas e tendências das aplicações de Web Semântica no governo eletrônico brasileiro, destacando os aspectos de interoperabilidade semântica entre os processos organizacionais, com base nos padrões de interoperabilidade propostos pelo e-PING. O trabalho apresenta uma proposta de um modelo para interoperabilidade semântica envolvendo processos G2G no e-Gov brasileiro. Apresentam-se ainda algumas propostas de complementações às especificações do e-PING que suportem o modelo proposto. Por fim, um cenário de aplicação do modelo proposto é discutido. Palavras-Chave ⎯ padrões de interoperabilidade semântica, especificações do e-PING, web services semânticos. I. INTRODUÇÃO Nos últimos anos, as organizações corporativas e setores públicos de vários países têm participado de um conjunto de iniciativas de busca por inovações em sua estrutura organizacional de forma a viabilizar a entrega de serviços aos cidadãos em todo o mundo [1]. Nesse contexto, as iniciati...
Hoje, os sistemas de comando e controle são altamente dependentes da infraestrutura da informação... more Hoje, os sistemas de comando e controle são altamente dependentes da infraestrutura da informação (TI). Este fato torna um desafio para os comandantes identificar como os aspectos cibernéticos incluenciam na consecução dos objetivos relacionados com a missão. Apesar de sua importância, a compreensão do impacto dos aspectos cibernéticos no domínio físico (terra, ar e mar), ainda é algo novo e seus resultados ainda muito incipientes, especialmente quando se deseja compreendê-lo em tempo real. Este artigo apresenta os requisitos necessários para endereçar esta questão e apresenta um framework para realizar a avaliação do impacto cibernético em uma missão. Através deste framework, uma visão unificada de todos os aspectos da missão nos seus vários domínios é construída, permitindo que um analista priorize seus recursos e permitindo ao mesmo que o efeito desejado seja produzido
Fifth IEEE International Conference on Advanced Learning Technologies (ICALT'05), 2005
Modeling learner interaction in Educational Adaptive Hypermedia Systems (EAHS) is not normally dr... more Modeling learner interaction in Educational Adaptive Hypermedia Systems (EAHS) is not normally driven by the pedagogical model used in the system. It is argued in this paper that besides driving the modeling of learner interaction, the pedagogical model is a key aspect in modeling the system as a whole. This paper presents a proposal to define and present learner interaction based on Problem-based Learning. Though the proposed process has been used only in academic context, the preliminary use has indicated its potential of use in real contexts.
International Journal of Advanced Computer Science and Applications, 2012
There are many situations where it is needed to represent and analyze the concepts that describe ... more There are many situations where it is needed to represent and analyze the concepts that describe a document or a collection of documents. One of such situations is the information retrieval, which is becoming more complex by the growing number and variety of document types. One way to represent the concepts is through a formal structure using ontologies. Thus, this article presents a fast and simple method for automatic extraction of ontologies from documents or from a collections of documents that is independent of the document type and uses the junction of several theories and techniques, such as latent semantic for the extraction of initial concepts, Wordnet and similarity to obtain the correlation between the concepts.
Uploads
Papers by Jose Parente de Oliveira