Data Insecurity

In the modern age there are hundreds of companies that have access to its users personal information. Many companies require some sort of sign up process for first time users. This often involves the user giving the company their name, address, birthday, and in some cases even more sensitive information such as a credit card number. Most people do not expect this data to be leaked around the internet, but it happens quite often. When a data breach occurs, the sensitive information that is gathered is likely put up for sale on different dark web sites for anyone to purchase. This type of information breach can happen to any company, even if they might not seem like they would have much sensitive information on their customers. For example, Panera suffered a massive data breach last year where the data of 37 million customers were exposed. Here is a website that talks about some of the other biggest data breaches of last year, and dives more into the specifics of each breach. The cyber criminals that are responsible for these data breaches continue to find new hacking methods to uncover this data. So since any company seems to be susceptible to these data breaches, what can be done to improve data security and keep everyone’s data safe?

Big companies use a variety of techniques to keep their customer’s sensitive data safe. At a purely physical level they have many policies to restrict the possibility of a data breach. For example, many companies use encrypted hard drives to store information, encrypted USBs to protect moving data, and also encrypted phones to protect data shared through telephone. Many companies have policies that require theses devices to be used, and also extra policies about employees own devices. Employees are often required to use a laptop or other device that has no USB slots, and is not able to download or export data over cloud. This is to prevent data from reaching outside sources by employees. A statistic that I got from this website (also a very interesting article) says that ignorance and negligence from employees cause 54% of all data breaches.

Many people also falsely believe that the big companies just have all the data encrypted, so it would not be accessible anywayss, but that is often not the case. Most large amounts of data from companies get stored in a relational database, as it is the easiest method of storing big data. However, it is difficult to encrypt data that is stored in a relational database, so whoever has access to it can often just read the data inside. Encrypting a database is also very expensive when you are purchasing that database from another company. All companies should be required to encrypt their customers sensitive data. It is a major violation of data integrity.

People tend not to think very often about how much information about their personal lives are truly out there for companies to sell around. We have no real idea how well protected the data, that we casually enter in when registering for a website, truly is. Most likely it is going into a database with no real protection at all! Data security still has a long way to go, and certainly more companies need to start implementing better encryption of their customers data. We all need to be more careful with our sensitive information, and pause to think where exactly the credit card number we are entering in is really going.

Nick Bagley

Graph Overload

There are hundreds of different types of graphs that a person can use to represent data. This often makes it difficult to figure out which type of graph is the optimal choice to display the information most clearly. Different graphs are good for different purposes, and in this post I will discuss a few of the key graph types that can be used in common situations.

If the data that is being displayed is not overly complex, then often times the simplest graphs are the best to use. The basic bar graph is good to use when comparing different numerical values against each other. For example, if data is gathered on several group’s opinions on a topic, a bar graph is an easy way to represent the number of people from each group that favor one opinion or the other. This kind of graph is also very useful in a financial standpoint, allowing for different dollar amounts to be compared between different time periods or companies. Another very simple yet powerful graph is the line graph. This graph is mainly used to represent trends, which clearly shows if a certain data set is increasing or decreasing based on the parameters. One of the most recognizable uses of this graph is in representations of the stock market. It shows the trends of different stock prices, and allows the reader of the graph to very quickly identify which stocks have an upward trajectory and which do not. Both bar graphs and line graphs can be understood without much analysis, making them very useful for quick and easy representations of data.

While bar graphs and line graphs are very useful in industries such as politics and business, they are not as widely used in the more scientific fields. The kind of data gathered through scientific research does not always make sense when put into these graphs. This is where graphs such as scatter plots are useful. The scatter plot allows two variables to be considered, and when the points are analyzed a relationship between these two variables can be found. This is helpful for scientists to find patterns in their data and make new discoveries based on connections that would not be able to be seen otherwise. Spider charts are also very useful in the scientific world, allowing for more than two variables to be considered for the data. A single entry can be considered against multiple variables around the circle of the graph, and additional entries can be compared in the same graph by having a color key present. This allows scientists to identify the entries that are best suited for a specific variable, and which are far below the competition.

While many graphs purpose are to display data in the most efficient manner, there are also times where the goal of a graph might be to be understood as simply as possible to a large group of people. These are graphs that may want to be used for presentations in a meeting for a large number of people so that the main idea can be communicated clearly. A very strong example of a graph in this category is the pictograph. This is a graph where the data gathered is translated into pictures so that it can be easily visualized by the audience. Venn diagrams are another graph that audiences easily understand. The venn diagram clearly shows two opposing sides, and shows the audience where the two intersect and where they have their differences. While these types of graphs are not necessarily the best for representing complex data, or even simple data, they can have a strong effect on an audience because of how easy they are to understand, and how they do not force the audience to analyze raw data too intensely.

There are many more kinds of graphs that can be used in a number of different situations. While some are hard to read, and some have very specific usage, there is a graph for every data set. This website shows a large number of different charts and graphs, many of which I have never seen before. New graph types will be created constantly with all of the new types and representations of data being introduced in the modern world, and it is important to keep all of these graphs in your arsenal when dealing with the unavoidable mountain of information in today’s age.

Nick Bagley

Life in Diagrams

There are many complex processes that take place in the world, most of which would be incredibly hard to explain using only words. People have a hard time expressing what is inside their minds, often causing miscommunications when trying to explain their ideas. This causes the person who is listening to this idea to receive a less efficient version of the process than what the original person knows. Now when they go explain the process to someone else using only words, the same thing will happen and the listener will gain a less efficient version of the process. This will continue to happen, basically creating a giant game of telephone, until somewhere far down the chain the process is almost completely different than what the person who came up with it intended. This leaves all of the different members of the company with a different idea of how to complete a task, which can cause many problems for obvious reasons. It is crucial for all members of a company to follow the same processes so that their tasks get completed. This is where diagrams are an absolute necessity.

The flowchart began to be used heavily in the 1930’s. The first industry to adopt widespread use of these diagrams was the industrial engineering field. The diagrams depicted the steps of different engineering processes so that all people involved in the process could have the same understanding of what needed to be done. With a proper guideline to follow, the processes became much more efficient, because the diagram could be analyzed and made to follow the most efficient steps to take in the process. However, before computers, analyzing and altering massive industrial flowcharts was a huge process in itself. Some of the flowcharts that were made for complex processes would have hundreds of steps and connections. Since they were written out by hand at this point in time, adding in a step, or changing an existing step, could completely change the way all of the connections worked in the diagram. The entire flowchart would often have to be remade to accommodate one extra step, since erasing and rewriting hundreds of lines and boxes made the diagram extremely messy. Now with modern technology, complex diagrams are much easier to create, alter, and store. This allows companies to rely more heavily on flowcharts for processes, maximizing efficiency. Multiple virtual copies of these charts are stored in databases, not on paper in files like they used to be. This means different people in the company can access them to understand the process better, and the entire operation can run much smoother. This ease of creating diagrams in the modern age has caused process charts to extend outside of just commercial use, and into the daily lives of people.

There is some sort of diagram on the internet for just about any sort of process imaginable. Here is the link to a site that shows a list of comedic flowcharts used for all sorts of different purposes. While the flowcharts on the website are obviously not very practical and are meant more as a joke than anything else, they can still actually be followed and ultimately do work. This just shows that truly anything in the daily lives of people can be represented as a diagram. Simple processes often do not need a diagram to make them more efficient, but they can still be made. If a person were to have a diagram for every process they forego in their day, with each diagram having the most efficient steps for completing the process, then the person would likely complete everything that they would have accomplished in half the time. Humans by nature do not follow the most efficient way of doing things. We are certainly capable of finding the most efficient way, and then following the diagram that explains it, however for most things in life we do not analyze processes for efficiency. Mapping out everything that a person does in diagrams would certainly make them accomplish their tasks quicker, but there would be no spontaneity, and ultimately that person would likely feel less human. While diagrams for large processes are extremely helpful, perhaps they should only be made at this larger level, because planning out simple processes in a person’s live can start to make living itself feel like a process, and that is something we certainly want to avoid.

Nick Bagley

Genetic Databases

Databases are used to store all sorts of different information across many different fields. The health industry utilizes databases to solve a multitude of problems through research and analysis on the data. An example of a type of database that is used in the health industry is the genetic database. These are databases that contain information about an organism’s genes, DNA variants, and much more genetic information. By having access to thousands of different organism’s genetic information, scientists can determine patterns in DNA to conclude which genes might be responsible for certain effects in the organism. This method of analysis is helping scientists determine the cause of some of the world’s biggest biological problems, including the father of them all; cancer.

There is an extremely interesting project going on in the health industry called TACCO, which stands for Transcriptome Alterations in Cancer Omnibus. This is an example of a genetic database, containing information on altered cancer genes. Through analysis of this database, researchers can make conclusions about risk levels for different types of cancers in people. Through analyzing the empirical data stored in TACCO, scientists can come up with numbers for the risk probability of a person with certain genes developing a specific type of cancer. For example, if the database contains one thousand instances of a person with a modified gene type A, and ten of those people developed brain cancer, then a link between people of that gene type and brain cancer could be made. If a patient with the same modified gene type A is seen by a doctor, the doctor could use these statistics to determine that this person has a 1% chance of developing brain cancer. The data in this database can also be used to determine survival rates of people with certain types of cancer based on their genetic makeup. There might be an extremely low survival rate for a specific type of cancer, let’s say 2%. However, through analyzing the data in the database, researchers could find that every person who had gene type B survived this cancer. This means that if a person with this gene type is diagnosed with this specific cancer, there survival rate would be above the normal 2%. Even more crucial, this would allow scientists to study this specific gene and figure out what it is about it that allows the people who have it to survive. This research could ultimately result in a cure for that specific type of cancer.

The information in these genetic databases are looked at in different formats to determine patterns. While a pattern between a certain genome may not be evident when analyzing the database itself, it might surface when the database is represented graphically. The information needed to find the causes and cures of earth’s biggest diseases might very well be there in these already existing databases. Now researches and scientists must deeply analyze this data in different ways to extract the patterns hidden among the lines of data. We are getting closer everyday to reaching these cures, and information and data scientists are needed more than ever to find them.

Nick Bagley

Data Encroachment

Most people would say that privacy is a crucial right, however big companies and organizations, such as Google and the federal government, have been pushing the boundaries of privacy. The average person generates approximately a gigabyte of data everyday, and more of that data than ever before can now be collected and used in different ways. However, many people do not want this data to be collected from them. A recent example of data collection going too far is the case against Mark Zuckerberg and Facebook last year. Here is a link to the article on the case from the New York Times, including a video of Zuckerberg’s testimony. Many people are outraged at the fact that their data can be harvested and seen by huge companies, and often times sold to other companies so that your private information has spread to many different areas.

This invasion of people’s data has caused people to alter the way they use the internet. Most people will at the least have some sort of cyber security system on their browser, such as a firewall. Many people are intent on making sure their privacy is kept, which springs up products such as a tabs for devices that cover the camera so that nobody can access it. While many people do not think that so much personal data should be collected and seen by others, there is a strong counterargument. More collection of data means major possible improvements across many different fields.

The health industry is an example of where an increase in gathering data could help a multitude of people. With modern technology scientists are able to create a device that people can put on that tracks a person’s heart rate, blood pressure, and other vitals at all times. This device can warn a person if there is an issue such as an upcoming heart attack so that they have enough time to seek medical help before it actually occurs. This device would be constantly gathering data from a person’s body, which is exactly why many people would not want something like this hitting mainstream usage. The human body generates about two terabytes of data everyday, and this technology is able to collect all of that data. This means that there could be thousands of different people that all have information on things like a person’s heartbeat at all times. This is seen as an invasion of privacy to many people, which is why limiting the misuse of data is an issue that the government pays strong attention to.

Data collection is used by companies to make the lives of their customers better. Sites such as Google collect an incredible amount of data from users which allows them to provide relevant advertisements and accurately predict what the user is going to search based on their past. However, when does this begin to invade a person’s privacy? Google could gather the data from a hospital showing that a patient just had a child. Then they would use this information and advertise products for babies and new mothers to the patient. This sharing of information between different companies and across industries is where people believe their privacy is intruded upon. Companies continue to try and collect more data and push the boundaries of privacy, and this is where we will see a true data encroachment.

Nick Bagley

Python vs. R

Back in 1991, Guido Van Rossum introduced the world to his new programming language, Python. The language entered into mainstream usage quickly, but only several years later, creators Ross Ihaka and Robert Gentleman created another programming language, R. Since then, both languages have been used heavily in the data analysis field. But which language is better? Since 2013, Python is being used by nearly four times as many people as R. Python has the fourth most active usage on Github and Stackoverflow, while R lands at 15th. However, that does not necessarily mean that Python is a better language for data science.

R is completely centered around data and statistical analysis. Data can be analyzed in tables, and manipulated with simple strings of commands. R provides it’s users with a plethora of base functions to extract information from data sets, and by combining these simple functions it is easy to produce a more complex command. Typically, R is not taught as a first programming language because it is known to be more difficult than languages such as Python. However, once the basic syntax is understood, it is easy to dive into everything R can do.

Another advantage that R has over Python is it’s code repository. R has a massive availability of packages to install, all available at CRAN, the Comprehensive R Archive Network. Python has a similar repository called PyPi, but it is not as heavily contributed to. This wide selection of packages allows R to continue to grow, while Python does not focus as much on the usage of packages.

However, with all of these advantages, Python is beginning to raise in popularity, looking to overtake R. As seen in the diagram below, more people are switching over to Python than ever before.

The world is becoming a greater environment for engineering. It isn’t only computer scientists that know how to code now. People in all different fields use some sort of coding in their occupation. That is why the adaptability of Python is beginning to take precedence over the raw functionality of R. Python code is easy to read, which means that people in different parts of a business can understand the code, even with no real knowledge of computer science. Python can also combine data analysis with programming better than R can. Python is much more applicable to engineering purposes and development purposes than R is, and there is more development happening in the world than ever before. This is why Python is becoming dominant in the world of data science.

-Nick Bagley

Swimming in Data

The first true form of data storage for machine use are punch cards. They were invented in 1725 and continued to be used commonly for centuries. They were cards with holes in them, and these holes represented instructions for the machine to follow. The most common uses for punch cards early on were textile looms and self playing pianos. These punch cards were easily understood by both humans and machines. By reading the documentation on the punch card, people could easily understand what the holes were supposed to do. For people to share this data, there would simply be increased production of the same punch cards. People use replicas of the same data to make machines carry out the same tasks. There was no need to go to computer science school to understand how these punch cards operate. Data storage was at a very elementary level, and there were no special languages required to extract the data and use it somewhere else.

It was not until 1948 that the first instance of RAM was introduced. Frederick Williams was able to store 1024 bits of information digitally. Intel started to release the first computer chip in 1966 that stored 2000 bits of information. Soon after external hard drives began to be made. Floppy disks and hard disks were introduced. However, even though there was more data able to be stored than ever before, it still was not easily transferable. To access this data, the physical drive had to be present, and the machines had to have a drive to access the data. There was no cloud where all of the information went. The data storage continued to improve. SSD cards and flash drives were invented in the 2000s to continue to make more storage and smaller physical chips possible.

In 2006 the term “cloud” was finally introduced. More data was being produced than ever before. Here is an interesting website that has a lot of interesting statistics about the amazing increase in data. By 2020 1.7 megabytes of information will be generated every second. This is the point in which data becomes difficult to access and to transfer. With the large amount of programming languages used it becomes harder to write universal programs manipulating the data. This is where XML files are necessary. The method of using XPath to access data from XML files is crucial to the sharing of this data.

Before information was accessible to everyone through a cloud, there was not as strong of a need to digitally transfer data between databases. However, now that it is possible for multiple people to access the same data, the question of how to efficiently manage it becomes increasingly more prominent. Data can be lost in a data lake, making it very difficult for anybody to access if it must be used. That is why it is important to have some way of transferring data into an XML. Once the data is organized in an XML format, the universally known XPath language can be used to access certain pieces of the information. Before data is generated, the software that generates it should have a way of storing that data into an organized store. This way the data can be utilized across different languages and machines. More data is about to be introduced into the world than ever before, and we cannot let it become lost.

-Nick Bagley

Catastrophic Miscommunications

Miscommunications happen all the time. It is very easy for somebody to mishear another person and think they said something other than what was spoken. However, miscommunications do not always come from mishearing something, but also from misinterpreting what is said. When people communicate what they want, they often are not very specific because what they say may seem obvious to them, but the listener could take it another way. Miscommunications happen in very high-stakes situations too, not just in normal conversation. A very famous example is a 1999 Mars mission lead by NASA which blew up when it entered the atmosphere of Mars. This was due to a lack of understanding between scientists about which units to use when doing their calculations. This resulted in some numbers being in Newtons while others related to pounds of force, and this miscommunication caused billions of dollars to blow up.

There are other examples of catastrophic miscommunications throughout history. In 1854 a British brigade advanced in a suicidal charge towards a much larger Russian force. However, this was not what the British commander wanted, but there was a miscommunication of orders through the ranks. The commander wanted the brigade to make sure the Russians did not move their heavy artillery guns. However, the brigade thought the commander wanted them to charge and try and reach the Russians artillery, which was located on the other side of the Russian defenses. This miscommunication of orders resulted in the lives of hundreds of British soldiers being taken.

Another tragic miscommunication was between the US and Japan. In 1945, a declaration of surrender was made by the US. When reporters asked Japan’s prime minister about Japan’s decision, he responded with the Japanese equivalent of “no comment.” However, the United States mistranslated his statement into something more closely resembling “not worthy of comment” or “holding in silent contempt.” This miscommunication prompted the US to launch an atomic bomb on Hiroshima just 10 days later, making this translation known as one of the most tragic miscommunications in history.

In my IS 2000 class, we discuss the challenges of gathering information. Miscommunication is always one of the greatest obstacles in eliciting information. There are many reasons this could happen such as a language barrier, a mishearing, or someone having trouble communicating what is in their mind. This is one reason why ontologies are incredibly helpful. They allow people to visualize what is going on in a system rather than just rely on implicit information and word of mouth. There is no language barrier when dealing with shapes and symbols on a chart, and no chance to mishear what the chart is saying. Working together to form one collective understanding of what is happening is necessary to successfully complete objectives. If the scientists at NASA had collectively decided on what units to use, then their shuttle would not have exploded. If the British had formed a battle plan on paper and made sure all commanders understood the plan before executing it then they would not have charged into certain death. A confirmation of Japan’s intentions or a better translation of the prime minister’s statement could have possible avoided Hiroshima. Miscommunications can be incredibly small yet still have enormous consequences. This is why information must be stored and organized before it is used, because a misinterpretation can be the cause of major problems.

Here is the link to an interesting article that talks about other major miscommunications in history. It is interesting to think about how the implementation of some sort of ontology could have prevented these events.

Nick Bagley

Object Oriented Ontology

This semester I am taking a computer science course that revolves around coding in Java. This is an object oriented language, and in just two weeks of using this language I can already see the importance of an ontology when coding in Java. In this language there is a type of data called an interface. This is a more broad data type in which more specific types of data called a class make up the interface. An example of something that we represented using an interface is the MBTA lines. The interface was the MBTA, and the classes that made up the interface were the red line, green line, orange line, blue line, and commuter rail.

When representing this data within Java, our instructors often suggest that we write a “class diagram.” This is a chart that shows which classes implement which interfaces, and how the different types of data relate to each other. The class diagram is an example of an ontology, and it closely resembles the UML diagram that we are learning about in my information science class. The data types are represented in boxes, and lines are drawn between the different boxes to represent the relationship that they have to each other. Inside each box in the class diagram are the fields that each type of data contain. These fields are equivalent to attributes in the UML diagram, and the data types are equivalent to entities.

In Java, when a method is applied to an interface, it then must be applied to all of the classes that make up that interface. This is an instance of inheritance, which is a characteristic of ontologies. Since the interface has a certain method, and the class is a kind of the interface, then the class must also have that method implemented. Transitivity is also found in the class diagrams that we create in Java. If x is a kind of y, and y is a kind of z, then x is a kind of z. This is a fundamental concept of logical reasoning in ontologies.

Creating class diagrams to represent data in Java is an example of where I use an ontology to simplify a task. Having the ontology available to reference makes it easier to understand which classes must implement a method when I add it to an interface. It allows me to quickly see what attributes each class has, and the relationships that they share with each other. Without understanding the taxonomy and partonomy that the data has, it is easy to make a mistake when coding that causes an error to occur in the program. Ontologies are crucial to use when coding in an object oriented language, especially when the relationships between the data represented become increasingly complex.

-Nick Bagley

What is information?

   “Information is a difference that makes a difference” – Gregory Bateson 

      When beginning the study of Information Science, it was evident very early on that there was a discrepancy over what information truly is. There are countless different definitions of what constitutes information, with every scholar, professor, and student having a slightly different outlook on the subject. The Wikipedia definition of information is something that “provides the answer to a question of some kind or resolves uncertainty.” This allows a very broad interpretation of what information is. However, experts including Theirauf and Floridi say that information can only be structured data, which eliminates any sort of unstructured data that can answer questions or resolve uncertainties. From reading all of these sources, I have generated my own idea of what information is.

     In our class, a definition for information was given that said information can be used to make decisions. This is similar to Wikipedia’s explanation about how information can answer questions. We also said in class that a tweet was not considered information in Floridi’s map because it was not structured data. However, I believe that a tweet can be used as information in other definitions. If somebody were to tweet something such as “I love the Patriots, Tom Brady is the best of all time” then that would answer the question of what football team that person supports. Tweets also cause many people to make decisions. For example, Donald Trump’s tweets cause many people to decide that he is not fit for presidency. People absorb these tweets as information to make decisions and resolve uncertainties they might have. In that sense, I believe that tweets should be classified as information. 

     The extent to what I think could be considered information goes on from just tweets. In one of the readings, a study showed that Americans spend half of each day consuming information. The study focuses on television and radio heavily, but I believe that information is consumed in many more places, even when somebody sleeps. If somebody were to watch a horror movie before bed, and then have a bad dream because of it, they might learn from it and decide not to watch horror movies before bed anymore. The dream was an event that happened that led a person to make an informative decision. Looking at it in that sense, that dream could be information, although many sources such as Floridi would not classify it as such because it is not structured. These sort of events happen constantly throughout the day, often subconsciously so that people do not even know they are absorbing information. People’s actions and decisions stem from information they are consuming, even if they do not realize they are consuming it. 

     Our professor told us that if you asked anybody in the field of information science what information is, they would give you a slightly different answer. Conforming to this statement, I have already begun to have my own idea of what information is, which is sure to change over time and over the course of this class. I believe that information is anything that can be used to alter somebody’s decisions, opinions, actions, or knowledge. This definition allows for a much broader classification of what is information. This idea of information follows most closely with Bateson’s interpretation, “information is a difference that makes a difference.”

–  Nick Bagley

Design a site like this with WordPress.com
Get started