• OBJECT ORIENTED ARTIFICIAL NEURAL NETWORK SIMULATOR IN TEXT AND SYMBOL RECOGNITION

      Piszcz, Alan; Ishaq, Naseem; Advisor; Novillo, Jorge E; Reviewer; Sengupta, Saumendra; Reviewer (1993)
      Objected oriented languages and artificial neural networks are new areas of research and development. This thesis investigates the application of artificial neural networks using an object oriented C++ backpropagation simulator. The application domain investigated is hand printed text and engineering symbol recognition. An object oriented approach to the simulator allows other simulator paradigms to reuse a large body of the object classes developed for this particular application. The review and implementation of image feature extraction methodologies is another area researched in this paper. Four feature techniques are researched, developed, applied and tested, using digits, upper case alphabet characters and engineering symbol images. Final implementation and testing of the feature extraction methods with a baseline technique is analyzed for applicability in the domain of hand printed text and engineering symbols
    • Aligning the SUNY Poly NCS Program with Nationally Recognized Accreditation

      Cook, John; Marsh, John; Adviser; Hash, Larry; Reviewer; Bull, Ronny; Reviewer (2015-01-29)
      This document is an exploration into what types of curriculum changes must be made to accommodate accreditation. In the review of program accrediting bodies, none is more authoritative or more appropriate than the Accreditation Board for Engineering and Technology (ABET). In ABET’s requirements for accreditation, computing related programs are defined and delineated. On further exploration, it can be seen that the Association for Computing Machinery (ACM) has driven the development of those definitions. The ACM further defines goals and objectives for these disciplines, as well as curriculum models. When reviewing other accreditations, not only are these ACM definitions recognized within those accreditations, goal and outcome alignment is also present. This ‘goal and outcome’ methodology is also present in institution level accreditations that SUNY Poly must comply with. After reviewing the ACM program definitions and comparing them to the NCS program, it is concluded that NCS most closely resembles an ACM IT defined program. This leads to the recommendation of adopting and aligning with ACM IT program guidelines, which provides solutions to multiple program and institution requirements as well as creating a solid pathway to accreditation.
    • Data Mining: Privacy Preservation in Data Mining Using Perturbation Techniques

      Patel, Nikunjkumar; Sengupta, Sam; Adviser; Andriamanalimanana, Bruno; Reviewer; Novillo, Jorge; Reviewer (2015-05-06)
      In recent years, data mining has become important player in determining future business strategies. Data mining helps identifying patterns and trends from large amount of data, which can be used for reducing cost, increasing revenue and many more. With increased use of various data mining technologies and larger storage devices, amount of data collected and stored is significantly increased. This data contains personal information like credit card details, contact and residential information, etc. All these reasons have made it inevitable to concentrate on privacy of the data. In order to alleviate privacy concerns, a number of techniques have recently been proposed to perform the data mining in privacy preserving way. This project briefs about various data mining models and explains in detail about perturbation techniques. Main objective of this project is to achieve two things. First, preserve the accuracy of the data mining models and second, preserve the privacy of the original data. The discussion about transformation invariant data mining models has shown that multiplicative perturbations can theoretically guarantee zero loss of accuracy for a number of models.
    • A Case Study on Apache HBase

      Nalla, Rohit Reddy; Sengupta, Sam; Adviser; Novillo, Jorge; Reviewer; Rezk, Mohamed; Reviewer (2015-05-16)
      Apache HBase is an open-source, non-relational and a distributed data base system built on top of HDFS (Hadoop Distributed File system). HBase was designed post Google’s Big table and it is written in Java. It was developed as a part of Apache’s Hadoop Project. It provides a kind of fault – tolerant mechanism to store minor amounts of non-zero items caught within large amounts of empty items. HBase is used when we require real-time read/write access to huge data bases. HBase project was started by the end of 2006 by Chad Walters and Jim Kellerman at Powerset.[2] The main purpose of HBase is to process large amounts of data. Mike Cafarella worked on code of the working system initially and later Jim Kellerman carried it to the next stage. HBase was first released as a part of Hadoop 0.15.0 in October 2007[2]. The project goal was holding of very large tables like billions of rows X millions of columns. In May 2010, HBase advanced to a major project and it became an Apache Top Level Project. Several applications like Adobe, Twitter, Yahoo, Trend Micro etc. use this data base. Social networking sites like Facebook have implemented its messenger application using HBase. This document helps us to understand how HBase works and how is it different from other data bases. This document highlights about the current challenges in data security and a couple of models have been proposed towards the security and levels of data access to overcome the challenges. This document also discusses the workload challenges and techniques to overcome. Also an overview has been given on how HBase has been implemented in real time application Facebook messenger app.
    • Spanning Tree Protocol

      Kasu, Sivalasya; Hash, Larry; Adviser; Marsh, John; Reviewer; Bull, Ronny; Reviewer (2015-05-21)
      This technology case study focuses on Spanning Tree Protocol (STP). The Spanning Tree Protocol is a layer2 protocol that ensures a loop free topology for any LAN network. The basic function of STP is to prevent loops in the network. Spanning Tree Protocol is standardized as IEEE 802.1D. Spanning tree is created within a Layer 2 network of connected switches, leaves only one active path between two network devices. Spanning tree has evolutions and extensions such as; Per VLAN Spanning Tree Protocol (PVST), Rapid Spanning Tree Protocol (RSTP) and Multi Spanning Tree Protocol (MSTP).
    • Representational State Transfer as a Web Service

      Desai, Dhruv; Sengupta, Sam; Adviser; Novillo, Jorge; Reviewer; Andriamanalimanana, Bruno; Reviewer (2015-12-01)
      This report is a study on Representational State Transfer architectural style and its usefulness for implementing web service. This report will highlight the differences in perceiving REST as an architectural style and as a web service. This document will also discuss web services in general and highlight important differences between the different web services in programming languages. The goal of this report is to clarify the term REST as an architectural style which has proved to be a popular choice for implementing a web service rather than REST being termed as a web service and compare Web Services based on its performance in a Java Application.
    • Enhancing the Effectiveness of Software Test Automation

      Jansing, David; Novillo, Jorge; Adviser; Cavallo, Roger; Reviewer; Spetka, Scott; Reviewer (2015-12-01)
      Effective software testing can save money and effort by catching problems before they make it very far through the software development process. It is known that the longer a defect remains undetected, the more expensive it is to fix. Testing is, therefore a critical part of the development process. It can also be expensive and labor intensive, particularly when done by hand. It is estimated that the total effort testing software consumes at least half of a project’s overall labor. Automation can make much of the testing an organization does more accurate and cheaper than merely putting several people in a room and having them run tests from a paper script. It also frees the testing staff to do more specific and in-­‐depth testing than would otherwise be possible. This paper focuses mainly on software test automation techniques and how automation can enhance the efficiency of a software team as well as the quality of the final product.
    • De-anonymizing Social Network Neighborhoods Using Auxiliary and Semantic Information

      Morgan, Steven Michael; Novillo, Jorge; Adviser; Andriamanalimanana, Bruno; Reviewer; Reale, Michael; Reviewer (2015-12-11)
      The increasing popularity of social networks and their progressively more robust uses provides an interesting intersection of data. Social graphs have been rigorously studied for de-anonymization. Users of social networks will provide feedback to pages of interest and will create a vibrant profile. In addition to user interests, textual analysis provides another feature set for users. The user profile can be viewed as a classical relational dataset in conjunction with graph data. This paper uses semantic information to improve the accuracy of de-anonymizing social network data.
    • Comparison of Network Switch Architectures by CISCO

      Vemula, Veera Venkata Satyanarayana; Hash, Larry; Advisor (2016-02-01)
      This project is targeted to compare two major switching architectures provided by CISCO. CISCO is a network device manufacturer who has contributed to networking world by inventing many networking protocols which are used to improve the network performance and network health. In this document the switching architectures CATALYST and NEXUS are compared. All the available features in each architectures are listed and working of the supported protocols is explained in detail. The document also considers three network scenarios and explains which architecture is best suited and explains why in detail.
    • Live Tweet Map with Sentimental Analysis

      Kotrika, Rohila; Chen-Fu Chiang; Reviewer; Saumendra, Sengupta; Advisor; Andriamanalimanana, Bruno; Reviewer (2016-05-01)
      This project basically aims to build a system for the real-time analysis of the trends and public views around the whole world by storing and analyzing the stream of tweets from the Twitter live API which produces a huge amount of data . The tweets, tweet ID, time and other relevant elements are stored into a database and are represented in a map that is being updated in near real time with the help of Google map API. This project also aims to achieve the sentimental analysis of the tweets by sending the tweets to the natural language processing API which in turn processes the tweets using the natural language processing and gives a result If those tweets are positive, negative or neutral in nature. The map clusters tweet as to show where people are tweeting most from according to the sample tweets we get from the streaming API. These clusters will be shown in different colors according to the sentimental evaluation we receive from the sentiment API by Vivek Narayanan which works by examining individual words and short sequences of words (n-grams) and comparing them with a probability model. The probability model is built on a pre labeled test set of IMDb movie reviews. It can also detect negations in phrases, i.e., the phrase "not bad" will be classified as positive despite having two individual words with a negative sentiment. The web service uses a co routine server based on event, so that the trained database can be loaded into shared memory for all requests, which makes it quite scalable and fast. The API is specified here, it supports batch calls so that network latency isn't the main bottleneck. For Instance, if a tweet is negative in evaluation then it is shown in a red color marker on the map, green for positive and grey for the neutral. This analytic will also demonstrate the heat map for all the tweets that are stored in the database which gives a satisfying answer demonstrating from which part of the world are most of the tweets from. In this project we create a dynamic web application with the target runtime environment as Apache Tomcat Server. The server will also be initialized with the context listener which starts running the code to get the tweets into the database till the server is stopped. The most popular trends among worldwide and citywide would be provided in a drop down to be selected from which gives a clear perspective on how each trend behaves. It also offers the public, the media, politicians and scholars a new and timely perspective on the dynamics of the world wide trends and public opinion.
    • High Performance Distributed Big File Cloud Storage

      Shakelli, Anusha; Sengupta, Sam; Adviser; White, Joshua; Reviewer (2016-05-01)
      Cloud storage services are growing at a fast rate and are emerging in data storage field. These services are used by people for backing up data, sharing file through social networks like Facebook [3], Zing Me [2]. Users will be able to upload data from computer, mobile or tablet and also download and share them to others. Thus, system load in cloud storage becomes huge. Nowadays, Cloud storage service has become a crucial requirement for many enterprises due to its features like cost saving, performance, security, flexibility. To design an efficient storage engine for cloud based systems, it is always required to deal with requirements like big file processing, lightweight metadata, deduplication, high scalability. Here we suggest a Big file cloud architecture to handle all problems in big file cloud system. Basically, here we propose to build a scalable distributed data cloud storage that supports big file with size up to several terabytes. In cloud storage, system load is usually heavy. Data deduplication to reduce wastage of storage space caused by storing same static data from different users. In order to solve the above problems, a common method used in Cloud storages, is by dividing big file into small blocks, storing them on disks and then dealing them using a metadata system [1], [6], [19], [20]. Current cloud storage services have a complex metadata system. Thereby, the space complexity of the metadata System is O(n) and it is not scalable for big file. In this research, a new big file cloud storage architecture and a better solution to reduce the space complexity of metadata is suggested.
    • Secure Routing in Mobile Ad-hoc Networks

      Jadhav, Sumedh; Hash, Larry; Adviser (2016-05-01)
      The detailed study investigates the various problems faced by Mobile Ad hoc Networks because of their inherent characteristics. The characteristics make these networks vulnerable to various attacks. The attack known to be performed on MANETs were researched in order to gain better understanding and insight on how they can be defended against. The focus of this project was to secure the routing protocols, since routing plays a vital role in the operation of network and is the target of most attacks. Various research papers were referenced for collecting the information needed. The existing routing protocols were systematically categorized and compared. Additionally, advanced routing protocols that integrate cryptographic methods into them were studied to come to a conclusion that they are the most robust protocols that effectively protect the Mobile Ad hoc Networks against some dangerous attacks. Securing the routing protocols is indeed the most important aspect for securing MANETs. However, there are other techniques that can be used as extensions for strengthening these networks, in addition to using a robust routing protocol. Recommendations have been made for the same.
    • Technology Case Study on Web Real-Time Communications (WebRTC)

      Karnati, Nagarjuna; Hash, Larry; Adviser (2016-05-01)
      Web real-time communication (WebRTC) is the latest technology standard which enables web browsers to communicate directly without having to install any internal or external plug-ins. WebRTC fills a critical gap in the web platform where a native proprietary app like Skype could do something which is media communication that World Wide Web just couldn’t. Now, this can be done form web using WebRTC technology. This paper starts with a brief introduction of WebRTC and how it got started. Moving on, it provides information about the WebRTC technical goals, architecture and protocols involved. This paper highlights the network address translation (NAT) traversal where STUN, TURN and ICE protocols are involved. Also, this paper highlights about the peer to peer to media flows with reference to WebRTC protocol stack and application program interface (API). In the end, this paper discusses about implemented security features, tools available for WebRTC development and provides enterprise use cases.
    • NautiCode: Coding for Kids

      Zeo, Brittany; Mullick, Rosemary; Adviser; Sarner, Ronald; Reviewer; Urban, Christopher; Reviewer (2016-05-08)
      Throughout my college career, I have asked students what made them decide to major in Computer Science. The answers I received very seldom revealed previous coding experience. Unfortunately, this is the system: you don’t know what you want to major in and so you choose something that looks interesting. You just hope it works out for the best. Fortunately for me, I had four years of programming experience in classes before reaching college as well as being a programmer on my high school’s FIRST Robotics team. This previous exposure to coding allowed me to make an educated decision about what I wanted to major in. It is not always the case that an individual gets this experience, and I want to change that. For my Masters Project, I have decided to come up with a website to get kids to learn and practice some basic concepts of coding: NautiCode. My target audience is mid to upper elementary school children. And best of all, there is no previous coding experience needed when using NautiCode. Even if Computer Science is not their career choice, they can have the exposure at an early age. Coding does not only benefit computer scientists; just having the background knowledge of concepts such as: logic, data storage, and how things relate, can be beneficial to an individual for any major. These ideas can help individuals think about problems differently and come up with solutions that would not have been possible had they not been exposed to computer science concepts. What better time in an individual’s life to introduce these concepts than childhood. Children’s brains are magnificent. They can absorb so much information and they think differently about the world. This leads to creative solutions and a new perspective. What I aim to do with NautiCode is to get children thinking in new ways and exploit their creativity and spark new ideas. I aim to give an explanation of the simple concepts in an introduction and gradually work up towards more difficult problems. Children are more capable than they know and with a little guidance, they can start creating their own technologies in no time. NautiCode is a fully functional website that I created on my own. The front end is using Scss, and HTML5 while the backend is using PHP, SQL, JS, and AJAX. My databases are being hosted locally through phpMyAdmin and MAMP.
    • Botnet Campaign Detection on Twitter

      Fields, Jeremy; Sengupta, Saumendra; Adviser; White, Joshura; Reviewer; Spetka, Scott; Reviewer (2016-08)
      The goal of this thesis is to investigate and analyze botnet activity on social media networks. We first start by creating an algorithm and scoring method for “likely bots,” and analyze them in conjunction with their neighboring messages to determine whether there is a likely group of bots, or botnet. Chapters 1 & 2 cover the overview of the work, and the previous research done by others. Multiple datasets were collected from Twitter, over different time frames, including random samples, and targeted topics. Chapters 3 & 4 cover the methodology and results of the approach using these datasets. The method is shown to have high accuracy.
    • Study of Brain Computer Interfacing (BCI) with Cochlear Implant as an Example

      Roy, Debsuta; Hash, Larry; Advisor (2016-08)
      Brain Computer interfacing (BCI) has been under constant research to improvise the pathway amongst the brain and outside subject in several characteristics. This technology is being widely appreciated as the world tends towards automation. The mechanics of this domain include computer prone brain interfaces induced on the cortical part of the brain which helps in controlling external devices via thought. These interfaces are designed such that they will receive and transmit acoustic data and process them so that they can be recorded as a generalization thereby helping in generating signals artificially. Cochlear Implant has been demonstrated as an example of BCI where stimulation of brain takes place with the help of an external device. As this field is still being developed, there are several factors needed to be focused viz. research/study of the subject, location of electrodes, cost, complexity, performance, setup requirements and security. BCI field is a highly vast topic and is a combination of various fields like science, technology, medicine etc. This piece of research and study focuses primarily on the neuroscience and technology part. This project is presented to showcase the depth of this field and create interest for further advancement in this field. Chapter one deals with introduction to the technology as a whole and its related study of papers. Chapter two discusses about brain and the principle behind this technology. Chapter three helps in understanding the technology by splitting its components. This is followed by taking an example of cochlear implant technology in chapter four. Chapter five initiates the wireless approach brought to this technology. And then chapter six is all about challenges, conclusions and ideas of enhancement.
    • Routing Protocols for MANETs

      Muchintala, Priyanka Reddy; Hash, Larry; Advisor (2016-08)
      This technology case study focuses on the routing protocols used in mobile ad-hoc networks. For any wireless ad-hoc network, routing is the main concept. Proactive and reactive routing protocols are two such approaches. The ever changing properties like asymmetric links, low transmission power and topology have always been a challenge for the routing protocols. Under such circumstances, the proactive and reactive routing protocols have proved to be inefficient. Each of these approaches has its own disadvantages. To overcome these disadvantages, hybrid routing protocol has been designed. This paper will provide an overview of all the three above mentioned protocols namely proactive routing protocol, reactive routing protocol and hybrid routing protocol with an example each. A comparison between the three routing protocols is also provided in this paper.
    • BGP Routing Protocol

      Parasa, Sai Kiran; Hash, Larry; Advisor (2016-08)
      Border Gateway Protocol is the protocol which makes the Internet work. It is used at the Service provider level which is between different Autonomous Systems (AS). An Autonomous System is a single organization which controls the administrative part of a network. Routing with in an Autonomous System is called as Intra-Autonomous routing and routing between different Autonomous Systems is called as Inter-Autonomous System routing. The routing protocols used within an Autonomous System are called Interior Gateway Protocols (IGP) and the protocols used between the Autonomous Systems are called Exterior Gateway Protocols. Routing Information Protocol (RIP), Open Short Path First (OSPF) and Enhanced Interior Gateway Routing Protocol (EIGRP) are the examples for IGP protocols and Border Gateway Protocol (BGP) is the example for EGP protocols. Every routing protocol use some metric to calculate the best path to transfer the routing information. BGP rather than using a particular metric, it uses BGP attributes to select the best path. Once it selects the best path, then it starts sending the updates in the network. Every router implementing BGP in the network, configures this best path in its Routing Information Base. Only one best route is selected and forwarded to the whole network. [17] Due to the tremendous increase in the size of the internet and its users, the convergence time during link failure in the protocol is very high.
    • A Technology Case Study on Integrating Open Stack with SDN for Internet Connectivity using BGP

      Gonuguntla, Raja Bhushan Rao; Hash, Larry; Advisor (2016-12)
      There were many developments in Internet usage, which resulted in significant increase in Internet routing. With existing networking infrastructure, it is difficult to meet these requirements and causing more cost and less performance. Since network devices are hardware modules, processing them requires more power and more memory. However, if network protocols are developed using software modules, flexibility can be achieved in various programming applications and reduces dependency on hardware. The concept of using networking protocols as a software module can be explained using “Software Defined Networking (SDN).” With SDN, existing infrastructure can be integrated with various applications and centralized control protocols can be developed. One of the key components of SDN is integrating with Cloud Computing, where many applications can be built, which can be used for on-demand services. Integrating cloud computing with SDN will create dynamic networks and reduces infrastructure costs. In this paper, a case was considered for providing better internet connectivity by building public & private networks using Open source cloud technology (OpenStack) and existing distribution environments. For connectivity, BGP was used as routing protocol as it is known to be well- suited for such environments. Both public and private networks were integrated with SDN for centralized control. OpenStack was used to build various network topologies using different plugins through SDN controller. This method allowed to develop SDN controller with global view of OpenStack networks. The same controller was connected to distributed layers using Open Flow protocol. Since, both OpenStack and distributed networks were attached to SDN controller, centralized control of network protocols could be achieved. This model of centralized networks could be very useful in reducing costs and improving network efficiency, especially in large scale deployments.
    • People’s Perception of Smart Home Devices in their Homes & the Factors which Influence this Perception

      Thakur, Versha; Hash, Larry; Adviser; Marsh, John; Reviewer; Tekeoglu, Ali; Reviewer (2016-12-01)
      This thesis is to determine people’s perception of Smart home devices and the factors which influence this perception. Smart Home devices are becoming more common as many companies are launching various such devices. Although many companies are pushing these devices with extensive marketing, do people feel a need for these devices in their home? This thesis tries to find the answer to this question and what factors are affecting these opinions. Focus group methodology is used for this thesis. It is commonly used to perform qualitative researches. Based on the analysis of the Focus Group discussions, it is found that most people (from the targeted population) will be interested in adopting these smart home devices. There are some concerns which will still need to be addressed by the company such as security, Privacy and some health concerns to make the consumers more comfortable in adopting these devices.