Enhanced Approach for the Classification of Ulcerative Colitis Severity in Colonoscopy Videos Using CNN

Access: Use of this item is restricted to the UNT Community
Ulcerative colitis (UC) is a chronic inflammatory disease characterized by periods of relapses and remissions affecting more than 500,000 people in the United States. To achieve the therapeutic goals of UC, which are to first induce and then maintain disease remission, doctors need to evaluate the severity of UC of a patient. However, it is very difficult to evaluate the severity of UC objectively because of non-uniform nature of symptoms and large variations in their patterns. To address this, in our previous works, we developed two different approaches in which one is using the image textures, and the other is using CNN (convolutional neural network) to measure and classify objectively the severity of UC presented in optical colonoscopy video frames. But, we found that the image texture based approach could not handle larger number of variations in their patterns, and the CNN based approach could not achieve very high accuracy. In this paper, we improve our CNN based approach in two ways to provide better accuracy for the classification. We add more thorough and essential preprocessing, and generate more classes to accommodate large variations in their patterns. The experimental results show that the proposed preprocessing can improve the overall accuracy …
Date: August 2019
Creator: Sure, Venkata Leela
System: The UNT Digital Library

Mining Biomedical Data for Hidden Relationship Discovery

Access: Use of this item is restricted to the UNT Community
With an ever-growing number of publications in the biomedical domain, it becomes likely that important implicit connections between individual concepts of biomedical knowledge are overlooked. Literature based discovery (LBD) is in practice for many years to identify plausible associations between previously unrelated concepts. In this paper, we present a new, completely automatic and interactive system that creates a graph-based knowledge base to capture multifaceted complex associations among biomedical concepts. For a given pair of input concepts, our system auto-generates a list of ranked subgraphs uncovering possible previously unnoticed associations based on context information. To rank these subgraphs, we implement a novel ranking method using the context information obtained by performing random walks on the graph. In addition, we enhance the system by training a Neural Network Classifier to output the likelihood of the two concepts being likely related, which provides better insights to the end user.
Date: August 2019
Creator: Dharmavaram, Sirisha
System: The UNT Digital Library

Parallel Analysis of Aspect-Based Sentiment Summarization from Online Big-Data

Access: Use of this item is restricted to the UNT Community
Consumer's opinions and sentiments on products can reflect the performance of products in general or in various aspects. Analyzing these data is becoming feasible, considering the availability of immense data and the power of natural language processing. However, retailers have not taken full advantage of online comments. This work is dedicated to a solution for automatically analyzing and summarizing these valuable data at both product and category levels. In this research, a system was developed to retrieve and analyze extensive data from public online resources. A parallel framework was created to make this system extensible and efficient. In this framework, a star topological network was adopted in which each computing unit was assigned to retrieve a fraction of data and to assess sentiment. Finally, the preprocessed data were collected and summarized by the central machine which generates the final result that can be rendered through a web interface. The system was designed to have sound performance, robustness, manageability, extensibility, and accuracy.
Date: May 2019
Creator: Wei, Jinliang
System: The UNT Digital Library

A Language and Visual Interface to Specify Complex Spatial Pattern Mining

Access: Use of this item is restricted to the UNT Community
The emerging interests in spatial pattern mining leads to the demand for a flexible spatial pattern mining language, on which easy to use and understand visual pattern language could be built. It is worthwhile to define a pattern mining language called LCSPM to allow users to specify complex spatial patterns. I describe a proposed pattern mining language in this paper. A visual interface which allows users to specify the patterns visually is developed. Visual pattern queries are translated into the LCSPM language by a parser and data mining process can be triggered afterwards. The visual language is based on and goes beyond the visual language proposed in literature. I implemented a prototype system based on the open source JUMP framework.
Date: December 2006
Creator: Li, Xiaohui
System: The UNT Digital Library

A Multi-Variate Analysis of SMTP Paths and Relays to Restrict Spam and Phishing Attacks in Emails

Access: Use of this item is restricted to the UNT Community
The classifier discussed in this thesis considers the path traversed by an email (instead of its content) and reputation of the relays, features inaccessible to spammers. Groups of spammers and individual behaviors of a spammer in a given domain were analyzed to yield association patterns, which were then used to identify similar spammers. Unsolicited and phishing emails were successfully isolated from legitimate emails, using analysis results. Spammers and phishers are also categorized into serial spammers/phishers, recent spammers/phishers, prospective spammers/phishers, and suspects. Legitimate emails and trusted domains are classified into socially close (family members, friends), socially distinct (strangers etc), and opt-outs (resolved false positives and false negatives). Overall this classifier resulted in far less false positives when compared to current filters like SpamAssassin, achieving a 98.65% precision, which is well comparable to the precisions achieved by SPF, DNSRBL blacklists.
Date: December 2006
Creator: Palla, Srikanth
System: The UNT Digital Library

A Netcentric Scientific Research Repository

Access: Use of this item is restricted to the UNT Community
The Internet and networks in general have become essential tools for disseminating in-formation. Search engines have become the predominant means of finding information on the Web and all other data repositories, including local resources. Domain scientists regularly acquire and analyze images generated by equipment such as microscopes and cameras, resulting in complex image files that need to be managed in a convenient manner. This type of integrated environment has been recently termed a netcentric sci-entific research repository. I developed a number of data manipulation tools that allow researchers to manage their information more effectively in a netcentric environment. The specific contributions are: (1) A unique interface for management of data including files and relational databases. A wrapper for relational databases was developed so that the data can be indexed and searched using traditional search engines. This approach allows data in databases to be searched with the same interface as other data. Fur-thermore, this approach makes it easier for scientists to work with their data if they are not familiar with SQL. (2) A Web services based architecture for integrating analysis op-erations into a repository. This technique allows the system to leverage the large num-ber of existing tools by wrapping them …
Date: December 2006
Creator: Harrington, Brian
System: The UNT Digital Library

Using Reinforcement Learning in Partial Order Plan Space

Access: Use of this item is restricted to the UNT Community
Partial order planning is an important approach that solves planning problems without completely specifying the orderings between the actions in the plan. This property provides greater flexibility in executing plans; hence making the partial order planners a preferred choice over other planning methodologies. However, in order to find partially ordered plans, partial order planners perform a search in plan space rather than in space of world states and an uninformed search in plan space leads to poor efficiency. In this thesis, I discuss applying a reinforcement learning method, called First-visit Monte Carlo method, to partial order planning in order to design agents which do not need any training data or heuristics but are still able to make informed decisions in plan space based on experience. Communicating effectively with the agent is crucial in reinforcement learning. I address how this task was accomplished in plan space and the results from an evaluation of a blocks world test bed.
Date: May 2006
Creator: Ceylan, Hakan
System: The UNT Digital Library

Automated Defense Against Worm Propagation.

Access: Use of this item is restricted to the UNT Community
Worms have caused significant destruction over the last few years. Network security elements such as firewalls, IDS, etc have been ineffective against worms. Some worms are so fast that a manual intervention is not possible. This brings in the need for a stronger security architecture which can automatically react to stop worm propagation. The method has to be signature independent so that it can stop new worms. In this thesis, an automated defense system (ADS) is developed to automate defense against worms and contain the worm to a level where manual intervention is possible. This is accomplished with a two level architecture with feedback at each level. The inner loop is based on control system theory and uses the properties of PID (proportional, integral and differential controller). The outer loop works at the network level and stops the worm to reach its spread saturation point. In our lab setup, we verified that with only inner loop active the worm was delayed, and with both loops active we were able to restrict the propagation to 10% of the targeted hosts. One concern for deployment of a worm containment mechanism was degradation of throughput for legitimate traffic. We found that with proper …
Date: December 2005
Creator: Patwardhan, Sudeep
System: The UNT Digital Library

Procedural content creation and technologies for 3D graphics applications and games.

Access: Use of this item is restricted to the UNT Community
The recent transformation of consumer graphics (CG) cards into powerful 3D rendering processors is due in large measure to the success of game developers in delivering mass market entertainment software that feature highly immersive and captivating virtual environments. Despite this success, 3D CG application development is becoming increasingly handicapped by the inability of traditional content creation methods to keep up with the demand for content. The term content is used here to refer to any data operated on by application code that is meant for viewing, including 3D models, textures, animation sequences and maps or other data-intensive descriptions of virtual environments. Traditionally, content has been handcrafted by humans. A serious problem facing the interactive graphics software development community is how to increase the rate at which content can be produced to keep up with the increasingly rapid pace at which software for interactive applications can now be developed. Research addressing this problem centers around procedural content creation systems. By moving away from purely human content creation toward systems in which humans play a substantially less time-intensive but no less creative part in the process, procedural content creation opens new doors. From a qualitative standpoint, these types of systems will not …
Date: May 2005
Creator: Roden, Timothy E.
System: The UNT Digital Library

Peptide-based hidden Markov model for peptide fingerprint mapping.

Access: Use of this item is restricted to the UNT Community
Peptide mass fingerprinting (PMF) was the first automated method for protein identification in proteomics, and it remains in common usage today because of its simplicity and the low equipment costs for generating fingerprints. However, one of the problems with PMF is its limited specificity and sensitivity in protein identification. Here I present a method that shows potential to significantly enhance the accuracy of peptide mass fingerprinting, using a machine learning approach based on a hidden Markov model (HMM). This method is applied to improve differentiation of real protein matches from those that occur by chance. The system was trained using 300 examples of combined real and false-positive protein identification results, and 10-fold cross-validation applied to assess model discrimination. The model can achieve 93% accuracy in distinguishing correct and real protein identification results versus false-positive matches. The receiver operating characteristic (ROC) curve area for the best model was 0.833.
Date: December 2004
Creator: Yang, Dongmei
System: The UNT Digital Library

Voting Operating System (VOS)

Access: Use of this item is restricted to the UNT Community
The electronic voting machine (EVM) plays a very important role in a country where government officials are elected into office. Throughout the world, a specific operating system that tends to the specific requirement of the EVM does not exist. Existing EVM technology depends upon the various operating systems currently available, thus ignoring the basic needs of the system. There is a compromise over the basic requirements in order to develop the systems on the basis on an already available operating system, thus having a lot of scope for error. It is necessary to know the specific details of the particular device for which the operating system is being developed. In this document, I evaluate existing EVMs and identify flaws and shortcomings. I propose a solution for a new operating system that meets the specific requirements of the EVM, calling it Voting Operating System (VOS, pronounced 'voice'). The identification technique can be simplified by using the fingerprint technology that determines the identity of a person based on two fingerprints. I also discuss the various parts of the operating system that have to be implemented that can tend to all the basic requirements of an EVM, including implementation of the memory manager, …
Date: December 2004
Creator: Venkatadusumelli, Kiran
System: The UNT Digital Library

Hopfield Networks as an Error Correcting Technique for Speech Recognition

Access: Use of this item is restricted to the UNT Community
I experimented with Hopfield networks in the context of a voice-based, query-answering system. Hopfield networks are used to store and retrieve patterns. I used this technique to store queries represented as natural language sentences and I evaluated the accuracy of the technique for error correction in a spoken question-answering dialog between a computer and a user. I show that the use of an auto-associative Hopfield network helps make the speech recognition system more fault tolerant. I also looked at the available encoding schemes to convert a natural language sentence into a pattern of zeroes and ones that can be stored in the Hopfield network reliably, and I suggest scalable data representations which allow storing a large number of queries.
Date: May 2004
Creator: Bireddy, Chakradhar
System: The UNT Digital Library

An Analysis of Motivational Cues in Virtual Environments.

Access: Use of this item is restricted to the UNT Community
Guiding navigation in virtual environments (VEs) is a challenging task. A key issue in the navigation of a virtual environment is to be able to strike a balance between the user's need to explore the environment freely and the designer's need to ensure that the user experiences all the important events in the VE. This thesis reports on a study aimed at comparing the effectiveness of various navigation cues that are used to motivate users towards a specific target location. The results of this study indicate some significant differences in how users responded to the various cues.
Date: December 2003
Creator: Voruganti, Lavanya
System: The UNT Digital Library

Improved Approximation Algorithms for Geometric Packing Problems With Experimental Evaluation

Access: Use of this item is restricted to the UNT Community
Geometric packing problems are NP-complete problems that arise in VLSI design. In this thesis, we present two novel algorithms using dynamic programming to compute exactly the maximum number of k x k squares of unit size that can be packed without overlap into a given n x m grid. The first algorithm was implemented and ran successfully on problems of large input up to 1,000,000 nodes for different values. A heuristic based on the second algorithm is implemented. This heuristic is fast in practice, but may not always be giving optimal times in theory. However, over a wide range of random data this version of the algorithm is giving very good solutions very fast and runs on problems of up to 100,000,000 nodes in a grid and different ranges for the variables. It is also shown that this version of algorithm is clearly superior to the first algorithm and has shown to be very efficient in practice.
Date: December 2003
Creator: Song, Yongqiang
System: The UNT Digital Library

Performance Evaluation of Data Integrity Mechanisms for Mobile Agents

Access: Use of this item is restricted to the UNT Community
With the growing popularity of e-commerce applications that use software agents, the protection of mobile agent data has become imperative. To that end, the performance of four methods that protect the data integrity of mobile agents is evaluated. The methods investigated include existing approaches known as the Partial Result Authentication Codes, Hash Chaining, and Set Authentication Code methods, and a technique of our own design, called the Modified Set Authentication Code method, which addresses the limitations of the Set Authentication Code method. The experiments were run using the DADS agent system (developed at the Network Research Laboratory at UNT), for which a Data Integrity Module was designed. The experimental results show that our Modified Set Authentication Code technique performed comparably to the Set Authentication Code method.
Date: December 2003
Creator: Gunupudi, Vandana
System: The UNT Digital Library

Self-Optimizing Dynamic Finite Functions

Access: Use of this item is restricted to the UNT Community
Finite functions (also called maps) are used to describe a number of key computations and storage mechanisms used in software and hardware interpreters. Their presence spread over various memory and speed hierarchies in hardware and through various optimization processes (algorithmic and compilation based) in software, suggests encapsulating dynamic size changes and representation optimizations in a unique abstraction to be used across traditional computation mechanisms. We developed a memory allocator for testing the finite functions. We have implemented some dynamic finite functions and performed certain experiments to see the performance speed of these finite functions. We have developed some simple but powerful application programming interfaces (API) for these finite functions.
Date: December 2003
Creator: Jeripothula, Ramesh
System: The UNT Digital Library

Web Services for Libraries

Access: Use of this item is restricted to the UNT Community
Library information systems use different software applications and automated systems to gain access to distributed information. Rapid application development, changes made to existing software applications and development of new software on different platforms can make it difficult for library information systems to interoperate. Web services are used to offer better information access and retrieval solutions and hence make it more cost effective for libraries. This research focuses on how web services are implemented with the standard protocols like SOAP, WSDL and UDDI using different programming languages and platforms to achieve interoperability for libraries. It also shows how libraries can make use of this new technology. Since web services built on different platforms can interact with each other, libraries can access information with more efficiency and flexibility.
Date: December 2003
Creator: Manikonda, Sunil Prasad
System: The UNT Digital Library

Bounded Dynamic Source Routing in Mobile Ad Hoc Networks

Access: Use of this item is restricted to the UNT Community
A mobile ad hoc network (MANET) is a collection of mobile platforms or nodes that come together to form a network capable of communicating with each other, without the help of a central controller. To avail the maximum potential of a MANET, it is of great importance to devise a routing scheme, which will optimize upon the performance of a MANET, given the high rate of random mobility of the nodes. In a MANET individual nodes perform the routing functions like route discovery, route maintenance and delivery of packets from one node to the other. Existing routing protocols flood the network with broadcasts of route discovery messages, while attempting to establish a route. This characteristic is instrumental in deteriorating the performance of a MANET, as resource overhead triggered by broadcasts is directly proportional to the size of the network. Bounded-dynamic source routing (B-DSR), is proposed to curb this multitude of superfluous broadcasts, thus enabling to reserve valuable resources like bandwidth and battery power. B-DSR establishes a bounded region in the network, only within which, transmissions of route discovery messages are processed and validated for establishing a route. All route discovery messages reaching outside of this bounded region are dropped, thus …
Date: August 2003
Creator: George, Glyco
System: The UNT Digital Library

Resource Allocation in Mobile and Wireless Networks

Access: Use of this item is restricted to the UNT Community
The resources (memory, power and bandwidth) are limited in wireless and mobile networks. Previous research has shown that the quality of service (QoS) of the mobile client can be improved through efficient resources management. This thesis contains two areas of research that are strongly interrelated. In the first area of research, we extended the MoSync Algorithm, a network application layer media synchronization algorithm, to allow play-out of multimedia packets by the base station upon the mobile client in a First-In-First-Out (FIFO), Highest-Priority-First (PQ), Weighted Fair-Queuing (WFQ) and Round-Robin (RR) order. In the second area of research, we make modifications to the DSR and TORA routing algorithms to make them energy aware routing protocols. Our research shows that the QoS of the mobile client can be drastically improved through effective resource allocation.
Date: August 2003
Creator: Owens, Harold, II
System: The UNT Digital Library

Routing Optimization in Wireless Ad Hoc and Wireless Sensor Networks

Access: Use of this item is restricted to the UNT Community
Wireless ad hoc networks are expected to play an important role in civilian and military settings where wireless access to wired backbone is either ineffective or impossible. Wireless sensor networks are effective in remote data acquisition. Congestion control and power consumption in wireless ad hoc networks have received a lot of attention in recent research. Several algorithms have been proposed to reduce congestion and power consumption in wireless ad hoc and sensor networks. In this thesis, we focus upon two schemes, which deal with congestion control and power consumption issues. This thesis consists of two parts. In the first part, we describe a randomization scheme for congestion control in dynamic source routing protocol, which we refer to as RDSR. We also study a randomization scheme for GDSR protocol, a GPS optimized variant of DSR. We discuss RDSR and RGDSR implementations and present extensive simulation experiments to study their performance. Our results indicate that both RGDSR and RDSR protocols outperform their non-randomized counterparts by decreasing the number of route query packets. Furthermore, a probabilistic congestion control scheme based on local tuning of routing protocol parameters is shown to be feasible. In the second part we present a simulation based performance study …
Date: August 2003
Creator: Joseph, Linus
System: The UNT Digital Library

Secret Key Agreement without Public-Key Cryptography

Access: Use of this item is restricted to the UNT Community
Secure communication is the primary challenge in today's information network. In this project an efficient secret key agreement protocol is described and analyzed along with the other existing protocols. We focus primarily on Leighton and Micali's secret-key agreement without the use of public-key encryption techniques. The Leighton-Micali protocol is extremely efficient when implemented in software and has significant advantages over existing systems like Kerberos. In this method the secret keys are agreed upon using a trusted third party known as the trusted agent. The trusted agent generates the keys and writes them to a public directory before it goes offline. The communicating entities can retrieve the keys either from the online trusted agent or from the public directory service and agree upon a symmetric-key without any public-key procedures. The principal advantage of this method is that the user verifies the authenticity of the trusted agent before using the keys generated by it. The Leighton-Micali scheme is not vulnerable to the present day attacks like fabrication, modification or denial of service etc. The Leighton-Micali protocol can be employed in real-time systems like smart cards. In addition to the security properties and the simplicity of the protocol, our experiments show that in …
Date: August 2003
Creator: Surapaneni, Smitha
System: The UNT Digital Library

Evaluation of MPLS Enabled Networks

Access: Use of this item is restricted to the UNT Community
Recent developments in the Internet have inspired a wide range of business and consumer applications. The deployment of multimedia-based services has driven the demand for increased and guaranteed bandwidth requirements over the network. The diverse requirements of the wide range of users demand differentiated classes of service and quality assurance. The new technology of Multi-protocol label switching (MPLS) has emerged as a high performance and reliable option to address these challenges apart from the additional features that were not addressed before. This problem in lieu of thesis describes how the new paradigm of MPLS is advantageous over the conventional architecture. The motivation for this paradigm is discussed in the first part, followed by a detailed description of this new architecture. The information flow, the underlying protocols and the MPLS extensions to some of the traditional protocols are then discussed followed by the description of the simulation. The simulation results are used to show the advantages of the proposed technology.
Date: May 2003
Creator: Ratnakaram, Archith
System: The UNT Digital Library

Study and Sample Implementation of the Secure Shell Protocol (SSH)

Access: Use of this item is restricted to the UNT Community
Security is one of the main concerns of users who need to connect to a remote computer for various purposes, such as checking e-mails or viewing files. However in today's computer networks, privacy, transmission to intended client is not guaranteed. If data is transmitted over the Internet or a local network as plain text it may be captured and viewed by anyone with little technical knowledge. This may include sensitive data such as passwords. Big businesses use firewalls, virtual private networks and encrypt their transmissions to counter this at high costs. Secure shell protocol (SSH) provides an answer to this. SSH is a software protocol for secure communication over an insecure network. SSH not only offers authentication of hosts but also encrypts the sessions between the client and the server and is transparent to the end user. This Problem in Lieu of Thesis makes a study of SSH and creates a sample secure client and server which follows SSH and examines its performance.
Date: May 2003
Creator: Subramanyam, Udayakiran
System: The UNT Digital Library

Automatic Software Test Data Generation

Access: Use of this item is restricted to the UNT Community
In software testing, it is often desirable to find test inputs that exercise specific program features. Finding these inputs manually, is extremely time consuming, especially, when the software being tested is complex. Therefore, there have been numerous attempts automate this process. Random test data generation consists of generating test inputs at random, in the hope that they will exercise the desired software features. Often the desired inputs must satisfy complex constraints, and this makes a random approach seem unlikely to succeed. In contrast, combinatorial optimization techniques, such as those using genetic algorithms, are meant to solve difficult problems involving simultaneous satisfaction of many constraints.
Date: December 2002
Creator: Munugala, Ajay Kumar
System: The UNT Digital Library