Project Abstract

Behavioral analysis of data in process systems is the ultimate application of my accumulated experiences in statistics,
mathematics, and physics. The processes themselves are derived from mathematical and physical models, and are streamlined
using optimization methods that are at the heart of calculus. As the theoretically formulated processes are turned into production,
statistical analysis is used to isolate and perfect different aspects of the procedures to ensure the utmost efficiency of the overall
system. Research in process engineering is essential to improving industrial practices. In the case of plastic injection molding,
inefficiencies in production process result in the disruption in the uniformity of the product, excess wastage of materials, and other
potential disasters related to inconsistent measurements and calibrations of machines. The use of data driven methods to analyze
the entirety of the production process assists with the perpetual task of improving and optimizing the outputs while reducing errors
and eliminating opportunities for failure.

Sunday, April 23, 2017

Weeks 4 and 5 at Microtech

As I had described in my last post about my internship, I was able to get right into my research project and apply my knowledge and understanding of statistics to industrial processes. Since then, I have expanded the scope of my work at Microtech even more. As I had described in an earlier post, one of the key technologies that we use is the Vision system, the system that uses cameras and a high speed computer to take measurements of parts passing through on a belt in order to determine if they fit the specifications, and shoots them off the track if they aren't. However, since the computer needs to maintain its speed and efficiency, it doesn't collect any of the data by itself. If desired, its possible to pull some data from the machine using a flash drive, but that isn't useful because its is an extremely inefficient process. If the data form the vision system could be collected, then the job of quality control becomes much easier because they aren't wasting time measuring parts that can be measured by a machine at a much faster rate and with a lower margin or variability. Data collection from the vision system would ultimately allow for a more automated production facility, helping to lower costs and increase both output and quality. There was a solution to this issue. Getting a dedicated server solves the issue of the extreme inefficiency, and we will be able to access all the data at any time. Since I have a pretty strong background in computers, I got into that project at Microtech, and since then have been working closely with many people across management to get this initiative running. First, we needed to purchase a server, so we spent a day shopping for servers with all the given specifications and requirements that the vision company requested. Then we had to set it up and get it online. This process took a few days, and I was surprised by the complexity of coordinating with a client company on the other side of the world. Finally we were able to get the server running the Vision server software, but we are still working to connect the individual vision systems that are on the manufacturing floor to the server network. That is the task I am currently coordinating, and will hopefully be able to move forward with next week.


Background Research: Part 2

As I had discussed earlier, I had the opportunity to access some industry level certification materials, and here is a continuation of my background research.

Continuing from last time, I have delved further into the Master Black Belt materials. This time I focused on the Design of Experiments, as it is the essence of the research I work on at Microtech. In engineering, experiments are used for various functions, including process characterization, optimization, evaluation of material properties, product design, and system tolerance determination, among others. Experiments in process engineering mainly are used to reduce time to develop new processes, improve the performance of existing processes, improve the reliability of products, and to evaluate and explore material design alternatives. In industry, the design of experiments involves a few key components: Randomization, Replication, and Blocking. Randomization is probably the most simple, requiring that all the trials are run in random order, helping to eliminate other potential factors that may not be accounted for. Replication is also fairly straightforward, focusing on improving precision of effect estimation and error by having a reasonable sample size. Blocking is a practice that exists solely to eliminate nuisance factors, conditions over which we have no control but still influence the outputs since it is basically impossible to run a full DOE under homogeneous conditions. Experiments are usually designed using some mathematical structure, and the most common one is the factorial design. The factorial experiment is one in which all possible combinations are tested. The factorial design is noted as 2k, where the k is the number of factors being tested. The outputs are usually recorded in two levels, low and high. This structure is the foundation for many more advanced industrial experiments. Analysis of the factorial design has a few steps. First is estimating the factor effects and formulating a model. When using a replication design, the model should account for all factors, but when using an unreplicating design, normal probability plots can be used to model the effects as well. Once a model is generated, it is put through a statistical testing process called ANOVA (Analysis of Variance). Similar to a t-test, ANOVA allows us to compare different sets of data to determine whether or not there is a statistical difference between any number of data sets, but unlike a t-test it is much stricter, allowing for less type 1 errors (false positive) and therefore more widely accepted in industry. After testing the model, it can be refined and retested as many times as needed in order to reach a statistically favorable model. The final result usually comes in the form of a fitted regression model, a concept that I explored in my last background research post. 

Another interesting part of the six sigma materials I read was about innovation, or as they referred to it, the successful exploitation of ideas. As everyone knows, invention is a unique and novel device or discovery, usually a breakthrough in science and technology. Innovation, on the other hand, involves invention but is often the merging of different ideas and moving them through development, commercialization, production, and practical use. Design of Experiments is considered the foundation for large scale innovation, as interactions between various factors are often the key to innovation. Other than DOE, there are many other methods of innovation, such as data mining, statistics, business analytics, and other mathematical techniques. 

I will continue talking about some other background research in another post soon. 

Weeks 2 and 3 at Microtech

I returned to Microtech a couple weeks ago, and have since worked on many more aspects of the production process.

I started by continuing to familiarize myself with the key functions and methods used at Microtech. In addition to the measuring and auditing parts that I had learned to to before I left, I learned about a few more important functions of the quality lab. First is a process called burst testing. In a battery, the inside is filled with paste, and as the cell uses power the electricity contained in the compound undergoes chemical changes, causing it to release gases. In order to prevent the battery from exploding from use, the battery seals must be breathable enough to allow the produced gases to escape. In order to ensure that the seals are breathable enough, we put them into a machine called a burst tester. The seals are capped using metal parts similar to those you would see on a battery, and then are taken into a sealed chamber in which they are pressurized until the plastic seals burst. The machine records the pressures at which each of the seals bursts, so that we can analyze any trends in the data to identify if there are defective cavities or groups. For each type of battery seal, there are established limits on what is considered an appropriate pressure for the seal to burst at. If the seal bursts too early, then that means that the battery will die out early when begin used because the seal would pop before the all of the battery acid is used. If the seal bursts at a pressure that exceeds the required specifications, then that means that the rest of the battery would break before the seal, allowing batter acid to pour out into the device using it, potentially causing more damage. The next major task that I got into in the quality lab is a process called gage R&R. Gage R&R is a widely used statistical method in industry to ensure two major components of any production process: Repeatability and Reproducibility. In a gage study, multiple operators (people) measure the saem set of parts using some technique (gage) in a random double blind setting. Once measurements of all the parts are taken by all the operators, the values are entered into Minitab statistical software and evaluated. The results show the variability in the system, and minitab isolates the various sources of variability in the system - operator error, instrumental error, methodological error, and production error. Using these results, it becomes possible to determine the viability and efficiency of a certain gage as a measurement method. The repeatability and reproducibility are also analyzed, helping to provide mathematical backing for the variation in measurements by a single operator or instrument and between operators or instruments, respectively. 

Outside the quality lab, I delved right into the core of my research project, which is the statistical analysis of processes using the eDart system to collect data from the injection molding machines.The eDart collects data directly from the mold on various parameters such as Injection Integral, Peak Pressure, Shot temperature, and more. In industrial production, the goal is to optimize the production process so that each and every parameter is fine tuned to the best possible setting for the best production of high quality outputs, and in order to find out exactly what those best fitting settings are, we use a process called DOE ( Design of Experiments). In a DOE, each of the desired parameters are manipulated and isolated in a matrix that can then be compared with the output data in order to determine which changes resulted in positive, neutral, or negative standards of quality. Additionally, the data from the eDart allows us to make a couple additional inferences. First, it allows us to correlate the parameters that we manipulate (mold temperature, barrel temperature, hold pressure, etc.) with other parameters that we do not manipulate but still change, that we call covariates (injection integral, peak shot stroke, injection pressure, etc.). Second, it allows us to see similarities between the covariates and the final quality of the products. By allowing us to further break down the process and to create these inferences, the task of creating the optimal process becomes easier and more concrete. 


Wednesday, March 8, 2017

Background Research: Part 1

Hey everyone!


I know its been a while; I have been travelling and even though I have been continually researching, I haven't had stable internet access and so haven't been able to write out a post. 

I began my background research with Mr. Clark's materials from his Master Black Belt Six Sigma Training and Certification Course at the ASU School of Engineering. As a short summary, six sigma is a set of techniques, methods, and standards that are used in Process development in industry and manufacturing, as well as in business processes. The name itself comes from a common statistical practice that I learned in AP Stats. For data to be able to be statistically significant, it has to meet a certain threshold of satisfaction to prove that the results weren't due to random chance and that the data can be used to reasonably back hypotheses and draw conclusions, and the most commonly used threshold is 99.7% certainty, a value which corresponds with ±3σ (3 standard deviations either above or below the default value), totally a range of 6 standard deviations, thus 6σ. Six Sigma training is widely recognized and used across many industries, and the principles and methods can be applied to resolve a broad range of issues. The Master Black Belt is the highest level of certification, and other levels exist such as the green and black belt, and the course itself can be very expensive, and I had the privilege of accessing all of the materials from Mr. Clark's certification courses, and I am studying all the sections related to the application of statistical practices, such as multivariate regression, logistic regression, and categorical data analysis.

Multiple regression analysis is used when there are multiple inputs, but still one output. In simple linear regression, we deal with an input and an output, usually referred to as X and Y. The model for simple linear regression is y=β0+β1 X+ ε, where β0 and β1 are constants and ε represents random error. In multiple regression, there are multiple inputs, labelled X1, X2, X3,... The equation for this y=β0+β1X1+ β2X2+ β3X3+⋯ + βkXk+ ε for a model with k regression factors. Anyone who has takes calculus will recognize that this looks extremely familiar to a Taylor series expansion, and that is actually where this is derived from. The remainder is ε, and each dependent variable(X) had a coefficient(β). This model also can account for interactions between variables, adjusting the model to y=β0+β1X1+ β2X2+ β12X1X2 +  ε , and by letting X3 = X1 + X2, the model for interactions can be written the same as the original form, y=β0+β1X1+ β2X2+ β3X3+ε.  When using multiple regression, the goal is to create a fitted regression line using the above model that gives the best possible predictions, explains the behavior of the data, and uses as few independent variables as possible while providing good predictions for the response. The fitted regression line takes the form y ̂  = b0 + b1X1 + b2X2 + b3X3 .... bkXk. After developing the fitted regression line, it is subject to statistical testing to determine whether or not the independent variables are significantly related to the response variable. It is also subject to residual analysis and other model fitting methods to ensure that the line is the best possible line that accounts for the most possible data with the least residuals (difference between the actual data points and the value predicted by the line). Once the line passes all of these statistical examinations, it can be used to predict and model behavior the behavior of a certain output by manipulating the various inputs. 

I will continue talking about what I have learned from the six sigma materials as well as the other materials I borrowed from Mr. Clark in another post soon. 

Wednesday, March 1, 2017

First week at Microtech Southwest

February 13th was my first day at Microtech Southwest, the plastic injection molding facility where I am doing my project. The first week was all about getting to know all the parts of the fabrication process, as well as familiarizing myself with the various tasks and duties that needed to sustain production. I spent most of my time the first week in the quality lab, where finished parts are sorted, measured, tested, and verified to ensure that each batch of battery seals confirms to all required standards. from each batch of battery seals produced, one seal from each mold is taken out. That set of seals is measured for height, width, hub height, hub depth, and potentially other factors, depending on the model of the seal. One set of seals is also put through a burst tester, to see test the pressure tolerance of the seals. In a battery, the aqueous compound inside the cell can expand with use, and if the seal is breathable enough to maintain equilibrium inside the cell while allowing enough air through so that the cell doesn't explode, then it is perfect. The burst tester applies air pressure to the seals to see if they will fail under the similar conditions, and the data from the burst machine is yet another part of the quality assurance process. The last step of the quality assurance process is the statistical analysis of the data received from all the measurement machines. If the data is within an acceptable margin of error, the seals from that batch are accepted and can then be shipped off to the battery companies where the seals are then used to build batteries.

I also spent some time getting to know the other parts of the factory, mainly one of the vision systems. The vision system is a computer with several cameras connected to it. There is also a belt on which parts are fitted on by a machine, and the parts are passed through the cameras to measure similar values such as height, width, circumference, etc. The computer then processes the data from the cameras, and if the measured values are unacceptable the parts are knocked off the belt into a scrap pile. This machine operates at an extremely fast pace, with the ability to process up to 2000 parts per minute. This means that maintenance of the machine is of utmost importance because one error can cause hundreds of parts to be incorrectly sorted, which could mean that many good parts are rejected, wasting materials or that many bad parts are accepted, destroying the established quality standards. 

Recently, I have been away travelling, and due to the lack of stable internet access I haven't been able to post anything for the past couple weeks. During this temporary break from my time at Microtech, I have been doing some background research on both the plastic injection molding process and on the industrial statistical process, using materials borrowed from Mr. Clark, the director of Engineering at Microtech. I will follow up with a reflection on my findings in another post within the next week. 


Wednesday, February 8, 2017

Introduction

Process engineering involves the use of applied mathematics and statistics to improve industrial processes, and by using the eDart system I will be able to acquire data that can assist in the constant improvement and optimization of the plastic injection molding production process.

Behavioral analysis of data in process systems is the ultimate application of my accumulated experiences in high school statistics, mathematics, and physics. The processes themselves are derived from mathematical and physical models, and are streamlined using optimization methods that are at the heart of calculus. As the theoretically formulated processes are turned into production, statistical analysis is used to isolate and perfect different aspects of the procedures to ensure the utmost efficiency of the overall system.

Research in process engineering is essential to improving industrial practices. In the case of plastic injection molding, inefficiencies the production process result in the disruption in the uniformity of the product, excess wastage of materials, and other potential disasters related to inconsistent measurements and calibrations of machines.

The use of data driven methods to analyze the entirety of the production process assists with the perpetual task of improving and optimizing the outputs while reducing error and eliminating opportunities for failure.

I will use the eDart monitoring system to collect data about the plastic injection molding process. The eDart system will provide both the data needed to monitor production and to eliminate error in the system. The collected data will be analyzed using statistical software such as Minitab in order to draw conclusions about the efficiency and optimal performance of the process system. Depending of the results of each analysis, additional steps may be taken to collect more information or improvements can be made to specific aspects of the injection molding process.