University
stringclasses
19 values
Text
stringlengths
458
20.7k
Chalmers University of Technology
4. Process mining Figure 4.16: The process model of UI. As shown in the both figures, the green points are the start place T , and the red one I is the end place T . The blue boxes are the transitions, and the deeper the color, O the higher frequency they have. Although the two models are enough for further research (for example, machine learning), one cannot intuitively find the difference betweentwomodels. Ifwefilterthesmallprobabilityeventsandarcs, moreintuitive models which can represent the main flow of the events hidden behind the rather complex processes of figures 4.15 and 4.16, as illustrated in Figure 4.17. Compared to the BASE, participants would do something (release the gas pedal) before the emergency event happen. This confirmed that 3DATIS do made the participants’ behavior change, and the most important was that it made participants behaviors safer, i.e. they slow down before the dangerous event happened. Furthermore, with 3DATIS, the Accelerate happens before turning the steering wheel. On the con- trary, Accelerate happens after turning the steering wheel. It is probably because with 3DATIS, the speed has already slowed down, drivers need to accelerate in order to overtake the obstacle whereas without 3DATIS, drivers need to avoid the obstacle first, and then overtake. 43
Chalmers University of Technology
5 Conclusion In this thesis work, three main tasks for 3DATIS have been done: design, test and analysis. In the design task, the main question was solved: how to present the sound in order to make drivers easily understand the surrounding traffic informa- tion. Then, we tested 3DATIS with 30 participants and each of them did two rounds of experiments, one with 3DATIS and one without. Two groups of behavior data were collected from the experiments. When it came to data analysis, the driv- ing behavior of each scenario was studied. The performance of 3DATIS in Cutin, Pedestrian, Overtake and Intersection was fairly good. Especially in Intersection and Pedestrian, 3DATIS gave participants very positive indications to reduce the rate of collision and improve the TTC. However, in Redcab, except TTC (Time to collision), all the other indices went in the wrong direction. Most participants were not prepared and performed in panic with 3DATIS. As there were two obstacles in this scenario and one of them was the disturbance, participants could not distin- guish where the sound came from and thus their judgements were interfered. Also according to the questionnaire and interview, 3DATIS was sometimes annoy when there are more than one other road user in the road at the same time. Therefore, 3DATIS can hardly exert its advantages in a relatively sophisticated traffic envi- ronment and even disturbed participants’ judgement. In Chapter 4, a rather recent method called process mining to discover the process behind a given data set was introduced. Through process mining, the process models of scenario Cutin was dis- covered. By comparing the models with and without 3DATIS, the driving behavior with the help of 3DATIS was found to make participants behaviors safer. The limitation of this thesis work is mainly focused on the experiment. First, there were 30 participants that took part in the experiment and this number is almost the baseline that an experiment requires. With more participants, the conclusion would be more robust and reliable. Secondly, as each participant was required to finish two experiments with the same scenarios, the learning effect cannot be totally avoided, even though the randomization of scenarios can theoretically eliminate the bias. Thirdly, the steering wheel and pedals of the simulator was designed for the game, so it could make participants feel not like driving in a real world. The con- sequence of which is that collected data might not reflect the real driving habit of participants. 45
Chalmers University of Technology
Sensor calibration for autonomous mine vehicles Calibration of laser scanners and hinge angle sensor MAGNUS OLOFSSON Department of Applied Information Technology Chalmers University of Technology Summary Atlas Copco has delivered mining products since the last century and in 2007 they introduced an autonomous centre articulated LHD (Load-Haul-Dumper). LHD:s are used in most underground mines for transport of ore and the idea of an autonomous LHD was to improve safety, efficiency and productivity in underground mines. The autonomous LHD contains a number of sensors including an odometer, a hinge angle sensor in the articulation joint, a gyro and two laser scanners needed for autonomous operation. All of them need to be calibrated because of pose errors from high tolerances in the machine construction. In this thesis a method for calibrating the angular offset of the hinge angle sensor and another method for calibrating the angular offsets of the two laser scanners using data from the already existing sensors are presented. The hinge angle method relies on the gyro, the odometer and a kinematic model of a centre articulated vehicle to estimate the offset. For calibration of the laser scanners the offsets are augmented on to the states in the state space model used for positioning of the LHD. The augmented state is then estimated using a combination of an Extended Kalman filter and an Unscented Kalman filter in a SLAM algorithm that uses the kinematic model and the laser measurements. Experiments in simulated and real environments have shown that the hinge angle sensor estimates the angular offset to within 0.3o. Tests of the laser scanner calibration method have shown that it estimates the angular offset to within 0.5o. The methods run offline, are easy to operate and require that the operator drives at least 50 m in a straight mine drift at 2.5 m/s to collect enough data. The report is written in English. Keywords: calibration, offset, LHD, centre articulated, laser scanner, hinge angle, SLAM, Kalman filter. iii
Chalmers University of Technology
1 Introduction 1.1 History The mining industry is a dangerous workplace, especially underground mining. During the last decades mining companies have tried to improve efficiency, productivity and safety. Atlas Copco AB started out in 1873 as a railway company but turned into air compressors and air tools in the end of the 19th century. Today they are in a number of markets and at the beginning of the 20th century they started producing air driven rock drills to the mining industry. The development has moved on and today they produce a number of mining products including the centre articulated LHD (Load-Haul- Dump) Scooptram ST14 for loading, hauling and dumping materials in underground mines. All LHD:s have traditionally been manually driven by an operator placed in the machine. But for some years ago solutions began to be used where the LHD was remote controlled and teleoperated from a control room (Hainsworth 2001). With teleoperation the operator was removed from the dangerous mine, but teleoperation includes delays and the cameras restrict the view for the operator which decreases the productivity (Larsson, Broxvall & Saffiotti 2010). To solve the problems related to teleoperation and also remove the need of a operator to each LHD autonomous systems have been developed and are commercially available, see (Automine 2011), (Scooptram Automation 2011) and (MINEGEM 2011). Atlas Copco’s Scooptram Automation includes both teleoperation and autonomous operation. 1.2 System overview Atlas Copco’s autonomous system works using the principle Teaching, Route profiling and Playback. First the LHD is driven manually along the route it should operate on and all sensor data are logged. During route profiling data are converted to a number of grid maps (Marshall, Barfoot, & Larsson 2008) and a route profile containing waypoints, desired speeds and pause points for loading and dumping. During playback the route is followed with help of the sensors. Data from a hinge angle sensor in the articulation joint, a gyro and an odometer are weighted together using an EKF (Extended Kalman Filter) to estimate the position. Laser scans from two laser scanners are then compared with the map using UKF (Unscented Kalman Filter) to update the position estimate. Deviation from the desired route point is used as input to a control algorithm that steers the LHD towards the route point in the desired speed. The system is built to operate in an obstacle free environment, but simple obstacle avoidance algorithms are built in the system to make the machine stop if obstacles come in its way. 1
Chalmers University of Technology
1.3 Objective All of the sensors need calibration because of errors from high tolerances in the machine construction. This project focuses on calibration of the hinge angle sensor and the laser scanners. The hinge angle sensor is used to measure the angle between the front and rear part. A method was searched that could find an angle offset that when applied reduces it to within ±0.2o. Before this project it was required that someone climbed on the back of the LHD and tried to align the machines back- and front end while the LHD was moving. When aligned one was able to get a reading of the angular offset. To get correct maps it’s important that the orientation of laser scanners relative the LHD:s coordinate system is known. This project is focused on the laser heading, which is the most critical parameter. The goal was to find the angular offsets of the laser scanners to give a maximum scan matching error less than a predefined value of 0.35. Before this project the lasers heading where manually put to zero by eye sighting. 1.4 Contribution A new method is here proposed that can be used on any centre articulated vehicle to automatically find the offset of the hinge angle sensor. It uses the kinematic model in (Ridely & Corke 2001) together with the hinge angle sensor, the gyro and the odometer to get an estimate of the hinge angle offset. Error propagation was made to get an early estimate of the accuracy of the method. Different confidence intervals were then estimated to be able to guarantee its performance. Experiments have shown that the method delivers a hinge angle offset with an accuracy of ±0.3o. The requirement is that the machine is driven straight at least 33 m at 2.5 m/s A new method is here proposed that can be used to find the angular offset of a number of lasers on a moving vehicle automatically. It uses an augmentation of the original vehicle model, an EKF for prediction, an UKF for filtering and a SLAM (Simultanoues Localization and Mapping) algorithm. No error propagation nor a confidence interval were calculated for the laser offset because the method was too complex and too few data existed to get a good enough confidence interval. Experiments have verified that if the vehicle is driven straight in a 23 m normal sized mine drift it can estimate the angular offsets of two lasers to within ±0.5o, which is need to get a maximum scan matching error less than a predefined value of 0.35. 1.5 Report structure In Chapter 2 the report evaluates different ideas before going through the sensors, algorithms and methods used for finding the offsets of both the lasers and the hinge angle sensor in Chapter 3. After going through the methods it continues with dealing with how the driving and sensor measurements affect the accuracy of the estimated hinge angle in Chapter 4. It then goes through the real experiments done to verify the accuracy of the system and its results in Chapter 5. The report then proceeds with taking up the required angular 2
Chalmers University of Technology
2.2 Laser offsets A literature research showed that there didn’t exist any earlier work in the area of estimating the laser scanners angular offset. The closest work were the constant model error estimation in (Martinelli, Tomatis, Tapus, & Siegwart 2003) and (Madhavan, Dissanayake & Durrant-Whytel 1998) which uses EKF on laser data together with augmentation of the states. When estimating the laser offsets the laser measurements needs to be used, which increases the level of complexity compared to the hinge angle calibration case. One alternative was to by brute force test all positive combinations of laser offsets to see which two gave the smallest scan matching error but early experiments showed that it took too long time. The system is built around an EKF together with an UKF so a natural choice was to use them together with state augmentation to estimate the laser offset parameters. A problem with this solution was that the map needs to be known a priori, which isn’t always the case. To solve this problem we needed to implement SLAM. UKF based SLAM can be seen in (Martinez-Cantin & Castellanos 2005), which is one of many that uses one state for each feature. The original system uses a grid map and thereby the feature space become to large if each grid point should be represented by one state variable. A solution to this was created in (Merwe, Doucet, Freitas & Wan 2000) using a more complex particle filter. Instead of using it, it was concluded that one could create a less complex algorithm that reused the existing mapping algorithm together with the EKF and UKF. The computations of this method are still heavy and depend on a number of parameters but its faster than brute force search. 3 Theory and methods This chapter goes through the theory and methods used for estimating the hinge angle offset and the laser scanners angular offset. The chapter begins with going through the kinematic model of the centre articulated LHD in chapter 3.1 and then continues with the sensors used in Chapter 3.2. Chapter 3.3 goes through the hinge angle offset estimation and its error analysis together with the statistics used for calculation of the confidence intervals. Chapter 3.4 then continues with estimation of the laser scanners offset using EKF and UKF in a SLAM algorithm. The chapter ends with a brief walk-through of the implementation of the two methods and also the simulator used for creating simulated log-files. 4
Chalmers University of Technology
3.1 Kinematic model l F l R Figure 1: Model of a centre-articulated vehicle. The kinematic model of the LHD already in use in the system assumes no slip and can be found in a number of papers (Altafini 1999) and (Ridely & Corke 2001). It uses the setup seen in Figure 1 where g is the angle between the centre lines of the two bodies. q , q is the heading of the front and rear body R respectively in relation to the x-axis of the global coordinate system. l is the F distance between the hinge G and the front axle centre F and l the distance R between the hinge G and the rear axle centre R. The models for the rotation of the LHD are & vsing +l g& ( 1 ) q = R l cosg +l F R and & vsing - l g& cosg ( 2 ) q = F R l cosg +l F R where v is the speed of the front body. From the model it can be concluded that the rotation increases in magnitude with speed, hinge angle and if the hinge is rotatedg& „ 0. The rotation also decreases with increased machine length. The LHD has four wheel drive. There are separate differentials for the front and rear wheels but a stiff driveshaft connects between the two differentials. This makes the front and back wheels get the same speed. According to the model they shouldn’t have the same speed if l „ l or the vehicle is not driving F R straight,g „ 0, and thereby is a slippage introduced when g is increased. So the models performance decreases with increased g . In Ridely & Corke (2001) one error was found in Eq. ( 2 ) where they had missed the multiplication with cosg , which was proven by simple calculations and later confirmed by reading Altafini (1999). 5
Chalmers University of Technology
The nonlinear discrete time state space model in Marshall, Barfoot, & Larsson (2008) is created from a continuous one using Euler steps giving s(n+1) =Φ s(n)+Γ (s(n))(cid:215) (u(n)+v(n)), ( 3 ) s u x(n)   y(n) v(n) s(n) =   , u(n) =   q (n) w (n)   g(n) where  hcosq (n) 0    hsinq (n) 0   Γ (s(n)) =  - hsing(n) - hl  u R  l cosg(n)+l l cosg(n)+l   F R F R   0 h  v(n) is model errors and other uncertainties and assumed to be Gaussian noise v(n)~ Ν(0,Q) with covariance s 2 0  Q =  v   0 s 2  w h in the model is equal to the sampling time and w (n) is equal to the angular speed of the hinge angleg& (n). The state update matrix Φ is the identity matrix s because we don’t model the dynamics. 3.2 Sensors Figure 2: The LHD and its sensors. As seen in Figure 2 the LHD is equipped with a hinge angle sensor to measure the angle between the front and rear part. It uses and odometer mounted on the driveshaft in order to measure the speed and the distance travelled. Mounted on the top of the machine are two laser scanners used to map the surrounding and position the LHD. One of the scanners is mounted backwards on the rear part 6
Chalmers University of Technology
and one is mounted forward on the front part. Together with the front laser an IMU (Inertial Measurement Unit) is mounted, containing a heading gyro to measure the rotation of the front part. 3.2.1 IMU Navigation02 The IMU Navigation02 is delivered by the company AIMS. It has three gyros with 24 bits resolution each giving angular speed resolution of 16m o /s. The following parameters can be found for the gyros: • Range ±120 o/s • Bias Error 0.06 o/s (1 s ) • Scale Factor Error 0.25 % (1 s ) • Non-linearity 0.5 % of FS • Noise 0.1 o/s (Broadband RMS) • Bandwidth 25 Hz • Misalignment 5 mrad It’s only the heading gyro that is used and a simple model of its measurements is & & & q (n) =q (n)+q +v (n) m b q& m where v (n) ~ N(0,s 2 (n)) q& q& m m and & & s (n) =120*0.005/3+0.0025q (n)+0.001=0.3%+0.0025q (n). q& m & & q (n) is the true angular speed, q is the bias error and v (n) is the Gaussian b q& m noise affecting the measurements. The non-linearity and the scale factor are modelled as Gaussian noise even though they are constant for each unique IMU unit. Reasons for modelling them as Gaussian noise are that they aren’t known & and that they vary with the size of q so Gaussian noise is the closest b approximation. It’s approximated that three standard deviations contains all errors and that is why the non-linearity is divided by three to get one standard & deviation. The bias q is instead modelled as a constant because of that it’s b later estimated and removed in the analysis and its sample variance is small. Misalignment is approximated to be zero due to its small size. It should also be mentioned that the IMU is mounted in a rubber suspension to remove some vibrations. 7
Chalmers University of Technology
3.2.2 Odometer The odometer is base on a driveshaft encoder that measures the number of teeth passing by to calculate the distance travelled. It uses the fact that if the distance between the teeth, the radius of the wheels and the number of teeth passed are known the distance travelled d can be determined. The mounting on the m driveshaft makes the odometer insensible to turns. The odometer system consists of three parts, an encoder, an I/O module and the machine computer, which is illustrated in Figure 3. The I/O module checks the encoder pulse counter every 5 ms and the machine computer then checks the I/O module around every system sampling time h, equal with 40 ms. This makes it possible that the value in the I/O module is up to 5 ms old. When calculating the distance travelled the system calculate it as 40 ms has gone between two samples, which then isn’t always true. The error can never grow larger than a time error of around 5 ms. Machine computer I/O module Encoder 40 ms 5 ms pulse counter Figure 3: Odometer sample flow from sensor to machine computer The measured speed v is estimated from d using the model m m d (n)- d (n- 1) v(n)h– v(n)*5*10- 3 v(n)*5*10- 3 v (n) = m m = =v(n)– . m h h h It’s approximated that the synchronization error in each sample is zero mean Gaussian. According to the Gaussian distribution 3s » max error which v m makes v(n)*5*10- 3 s (n) = . v m 3h Drift due to slip is approximated to be zero according to previous measurements done with the machine. Also the radius of the wheels is said to be known without errors. 3.2.3 Hinge angle sensor The hinge angle sensor is an encoder mounted in the joint holding the front and rear body together. The encoders range is (0,360]o and it has 16 bits resolution giving an angular resolution of 6 millidegrees. Previous measures have shown that the angular measurements g (n) have a Gaussian noise v (n), which m g m have a standard deviation s of approximately 0.01o. The hinge angle sensor g m is mounted to give approximately 180o when the machine is straight. 0o is wanted so the offset g is 180o under ideal conditions but there is also a o mounting error that introduces an offset. The machine can operate in the range 8
Chalmers University of Technology
[-45,45]o where 0o is when the machine is straight. A measurement from the hinge angle can be modelled as g (n) =g(n)+g +v (n) ( 4 ) m o g m where g(n) is the true hinge angle and v (n)~ N(0,s 2 ). g g m m 3.2.4 SICK Laser scanner The SICK laser uses a moving laser beam to scan the environment in one plane. It has a maximum scanning range of 32 m in the interval [-90, 90]o degrees with an angular resolution of 1o giving N =181 scans. A laser beam i where points i=1,…, N can be modelled as points z (i) = z (i)+ z +v (i) m spot ci z where v (i) ~ N(0,s 2). z z z is a constant error in the range [0,0.03] m and s =0.01 m is the stochastic ci z error. The laser beam creates a spot, with aperture angle 0.11 milliradians, that increases with distance and angle to the objects normal. z (i) is the distance to spot the closest point on the object hit by the laser beam. 3.3 Hinge angle offset 3.3.1 Estimation of the hinge angle offset g o The idea is to rearrange Eq. ( 1 ) in Chapter 3.1 to get an expression for the hinge angle and then use it for estimation of the hinge angle offset g . o ˆ& Estimation is made by using the unbiased rotation speed q (n) measured by ubm the heading gyro, the measured speed v (n) and the hinge angle m measurementg (n), where n=1,…,N is the number of measurements. To m simplify the rearrangement estimation is done only for the rotation created from forward motion and not from change in hinge angle.g& is therefore put to zero and Eq. ( 1 ) is rearranged to give & & vsing - q (cid:215) l cosg =q (cid:215) l . F R It is known that adding two sinusoidal waves with the same frequency and different phase gives a new wave with the same frequency but a new phase and amplitude. For example asinx+bcosx = a2 +b2 sin(x+j ) 9
Chalmers University of Technology
where b 0 if a ‡ 0 j = tan- 1 +  a p if a <0 Applying this makes it possible to get only oneg in the equation v2 +( q& l ) 2 sin tan- 1- q& l F  +g  =q& l . F R     v     R Solving for g gives  &  & ( 5 )  ql  ql  g =sin- 1 R +tan- 1 F   ( )      v2 + q& l 2    v  R and for negative v  &  & ( 6 )  ql  ql  g =- sin- 1  R ( ) +tan- 1  F  .   v2 + q& l 2    v  R &ˆ The rotation q (n) includes both the rotation caused by the speed v(n)and by ubm g& (n). To remove the rotation caused by g& (n) we use the approximation g (n)- g (n- 1) gˆ& (n) » m m h which is independent of g and thereby only contains the articulation rate plus o some noise. The approximation is used in the non-speed part (v equal zero) of &ˆ Eq. ( 1 ) and is then subtracted from q (n) giving ubm & &ˆ l g&ˆ (n) ( 7 ) q (n) =q (n)- R . v ubm l cosgˆ(n)+l F R where gˆ(n) is an estimate of the true hinge angle. If g (n) is used as gˆ(n) it introduces an error in the final offset estimate m gˆ because of the offset. The error introduced is limited as long as g is o o reasonably close to zero but that isn’t always the case because the offset g can o be between (-180, 180]o. From the analysis in Chapter 4.2 it has been found that if the whole algorithm is iterated m=1,…,M times and gˆ(n) is estimated from g (n) by removing the previous offset estimate gˆ (m-1) the offset estimate m o gˆ (m) converges towards the true offset g . Given that the vehicle is driven o o along a straight line a first estimate of gˆ can be calculated from o 10
Chalmers University of Technology
3.3.3 Error propagation Using error propagation the propagation of the stochastic errors through the &ˆ hinge angle method can be analysed. By inserting combinations of q (n), ubm v (n), g (n) and the number of measurements N the variance of gˆ can be m m o found. It should be remembered that this requires that the model is correct and thereby doesn’t the analysis give any information regarding effects from errors in the model. In Eq. ( 7 ) the measurement error in all parameters are independent. The error propagation through Eq. ( 7 ) can then be estimated using & & & ¶ q ¶ q ¶ q s 2 = v s 2 + v s 2 + v s 2 q& &ˆ q& ¶ g g ¶ g g v ¶ q m m m m m- 1 ubm where & ¶ q v =1, &ˆ ¶ q ubm & ( ) ¶ q l +l cosg +l g - g sing v = - l h 2 1 m 1 m m- 1 m ¶ g 2 ( l +l cosg ) 2 m 2 1 m and & ¶ q l v = ( 2 ). ¶ g h l +l cosg m- 1 2 1 m In Eq. ( 5 ) the measurement errors are also independent giving ¶ g ¶ g s 2 = q& ,v s 2 + q& ,v s 2 g q&,v ¶ q& q& v ¶ v v m v m where ¶ g q& ,v = 1 l 2 v2 +q& v2l 12 - ( v2 +q& v2l 12) - 1 2l 12l 2q& v2 + l 1 ¶ q& v 1- q& v2l 22 v2 +q& v2l 12 v+q& v2l 12 & v2 +q 2l2 v v 1 and ¶ g q& ,v = - 1 q& vl 2v - l 1q& v . ¶ v m 1- q& v2l 22 ( v2 +q& 2l2) 23 v2 +q& v2l 12 & v 1 v2 +q 2l2 v 1 12
Chalmers University of Technology
The above equations holds for each measurement n=1,…,N. The total variance of the estimated offset gˆ is then from Eq. ( 10 ) o (( ) ) 1 N ( 11 ) sˆ2 = ∑ s 2 (n)+s 2 (n) +2r (n)s 2 (n)s 2 (n) . gˆ o N2 n=1 g q&,v g m g q&,v,g m g q&,v g m g is in Eq. ( 7 ) and in Eq. ( 9 ) which implies that there is a correlation m r . The variance s 2 is largest when r =1 but experiments have g q&,v,g m gˆ o g q&,v,g m shown that the effect of the correlation is small. Using sˆ one could estimate a 99 % confidence interval for the gˆ . The gˆ o o sensors have Gaussian errors and therefore the error on gˆ is also Gaussian. o Using the Gaussian distribution the confidence interval is given by gˆ - 2.576sˆ <g <gˆ +2.576sˆ o gˆ o o gˆ o o 3.3.4 Measurement statistics By repeating the same experiment N times and each time collect N 1 2 measurements the 99 % confidence interval of the hinge angle method can be found. There exist two methods to calculate the confidence interval. The first method begins with calculating the estimated standard deviation s (p), where p=1,…,N , of the sample offsets gˆ (n) using gˆ 1 so so s (p) = 1 ∑N 2 ( gˆ (p)- gˆ (n)) 2 gˆ so N - 1 o so 2 n=1 where n=1,…,N is the number of samples in each experiment. If N is large an 2 2 approximation can be made that s (p) is equal to the true standard deviation gˆ so s (p)of the whole population and thereby having a Gaussian distributed and gˆ so not a Student-t distribution. We assume that the spread of gˆ (p) is not correlated as a result of that the o measurement noise can be assumed to be white Gaussian noise. Then the estimated standard deviation of the gˆ (p) from all N measurements is given by o 2 1 ( 12 ) s (p)= s (p). gˆ 1 gˆ o N so 1 According to the Gaussian distribution one could know with 99 % confidence that the true offset g is in the interval o gˆ - 2.576s (p)<g <gˆ +2.576s (p). o gˆ 1 o o gˆ 1 o o This should be tested for all N experiments. One experiment had been enough 1 in the ideal world to assure the statistics but more experiments with the same 13
Chalmers University of Technology
setup are made to confirm the result. The interval should be smaller than ±0.2o to confirm that the method is good enough. Ideally it should be enough to increase N to get into that interval. Also if the spread is correlated s (p) is 2 gˆ 1 o going to be smaller than the true standard deviation. The second method calculates the standard deviation of gˆ (p) directly. To begin o with the mean of all gˆ (p) is o gˆ = 1 ∑N 1 gˆ (p) o N o 1 p=1 and the estimated standard deviation is then s = 1 ∑N 1 ( gˆ (p)- gˆ ) 2 . gˆ o N - 1 o o 1 p=1 Due to that N is relatively small s has an uncertainty. If the spread is 1 gˆ o Gaussian, s has a c 2 distribution and the 99 % confidence interval of s is gˆ gˆ o o (N - 1)s2 (N - 1)s2 L = 1 gˆ o <s < L = 1 gˆ o . 1 c 2 gˆ o2 2 c 2 0.99/2,N - 1 1- 0.99/2,N - 1 1 1 The confidence interval of g is then o gˆ - 2.576L <g <gˆ +2.576L o 2 o o 2 This method only relies on that the spread is Gaussian and is therefore more reliable. It gives a larger confidence interval due to the uncertainty in s . The gˆ 2 o uncertainty can be decreased by increasing N . When N goes to infinity the 1 1 confidence interval decreases to the true confidence interval. One could compare s with each of the N estimated standard deviations s gˆ 1 gˆ 1 o o by assuming that s is the true standard deviation and using the fact that s is gˆ gˆ o o the spread of the gˆ (p):s around gˆ . This gives o o c 2 = 1 ∑N 1 1 (gˆ (p)- gˆ )2. ( 13 ) N - 1 N - 1 s 2 o o 1 1 p=1 gˆo1 c 2 If >1 than s <s and we have under estimated the standard deviation N - 1 gˆ o1 gˆ o 1 using the first method. This means that the spread is probably not uncorrelated. c 2 can be use to scale s to its true value but it should be remembered N - 1 gˆ o1 1 that Eq. ( 13 ) doesn’t take in to consideration that s is an uncertain estimate gˆ o of s if N is small. gˆ 1 o 14
Chalmers University of Technology
3.4 Laser scanners angular offset 3.4.1 Creating and evaluating the map Under normal operation is the map M created under Route Profiling from the log data, containing all the sensor data recorded under the teaching step. To create M the log is gone through step by step to estimated u(n) using backward Euler ong and d . u(n) is then input to Eq. ( 3 ) to get the pose s(n) of the m m machine at each time step. The map, which is represented by a 2 dimensional binary grid is then finally updated using ray tracing along the recorded laser beams at each position s(n) for both the front and rear laser. After creation a simulated Playback is made in the map to evaluate it using the logged data as input. In the simulation the logged laser beams’ lengths are compared to the simulated ones. From the comparison over all samples a scan matching error e is calculated. If the map is identical to the real world the match machine should get identical laser scans and e equal to zero. The e match match must at all times be under 0.35 for the map to be acknowledged. 3.4.2 State estimation using nonlinear Kalman filter To keep track of the machine under playback the state s(n) needs to be estimated at each sample time. A good prediction of the state can be found by using the kinematic state space model in Eq. ( 3 ). To get better accuracy and compensate for drift the state is then filtered with help of the measured laser data. This method of first predicting using a model and then update using measured data is called Kalman Filtering. A standard Kalman Filter can only operate on linear equations. For nonlinear equations there exist a number of different solutions. This machine uses a mix of two Kalman Filters, an EKF (Extended Kalman Filter) for prediction and an UKF (Unscented Kalman Filter) for filtering. The implementation of the EKF and UKF used can be seen in (Marshall, Barfoot, & Larsson 2008). To simplify notation let sˆ- (n) be the prediction of the state using EKF and sˆ(n) the filtered estimate of the state using UKF. In the same way let P- (n) be the predicted state covariance and P(n)be the filtered covariance. To solve the problem that the Kalman filter needs linear equations the Extended Kalman Filter uses linearization of the state equation Eq. ( 3 ) around the previous state. Eq. ( 3 ) is linearized to get sˆ- (n) = H (sˆ(n- 1))(cid:215) sˆ(n- 1)+Γ (sˆ(n- 1))(cid:215) u(n). ( 14 ) s u 15
Chalmers University of Technology
s 2   xinit   s 2 0  yinit  s 2  P(0)=  q init .  s 2  g  0 init s 2   f  Rinit  s 2   f  Finit 3.4.4 SLAM The UKF needs a map to be able to estimate f and f but the mapping R F procedure in Chapter 3.4.1 is affected by f and f because of that can’t the R F map be created before the offsets are estimated. It needs to be done simultaneously. The map M could be created by adding laser scan information z(n) to M at each sample n using sˆ (n), estimated from Kalman filtering using artificial a ~ measurements Z(n) on M at n-1. This method has some drawbacks, one being ~ that earlier estimates of f and f affects Z(n) and experiments showed that R F this made the Kalman filter unable to make f and f converge to their true R F values. It’s solved by recreating the map around the machine at sample n from ~ scratch for each setup of f and f in S(n). To recreate the map ray tracing is R F used in each old pose in sˆ (n+n ) with scan z(n+n ), where a minus minus n = -n -2n ,…, -N and n is the updating step used to decrease minus step, step minus step computation time. Another drawback is that the map is less updated in the driving direction. This is solved by using the EKF to estimate future states sˆ- (n+n ), where a pozitive n =n 2n …, N The poses in sˆ- (n+n ) are then used positive step, step positive. a pozitive together with the laser scans z(n+n ) to add data to the recreated map for positive ~ each sigma setup in S(n) of f and f . This works over short travelled R F distances given that the model is correct. Each ray tracing to a point on the map overwrites the old value. So the update order is of importance. The start position of the machine is exactly known when mapping, because we decide its coordinates. Poses in negative time are also more accurate because of that they have been estimated using the whole Kalman filter. For this reason it is more convenient to start with mapping in the future poses and end with mapping in the negative poses. To get a more stable final output value the mean of the laser angle estimates fˆ R and fˆ in sˆ (n) from samples at n>N are used to get f and f . The first F a SM R F N samples is skipped to give the estimated offsets time to settle because it SM 19
Chalmers University of Technology
3.5 Implementation Laser and hinge angle offset estimation is made offline using a log-file containing all necessary data. The log file is recorded when the vehicle is driven a certain distance at a certain speed. To be able to create log-files, to test different laser offsets, an at Atlas Copco already existing SIMULINK simulator was modified. It simulated the kinematics of the ST14 seen in Eq. ( 3 ), the sensors and the laser scanning. Laser scans were originally made on maps created using lines but was modified to also handle grid maps. The laser simulation was also modified to include the laser spot effect discussed above so that the final simulator was able to simulate all errors in the laser sensor. Noises in the other sensors were skipped to decrease the complexity. The grid maps used were created using Profiling from real logs. The Route Profiling was together with the Simulated Playback, described in Chapter 3.4.1, already implemented by Atlas Copco in C++ and ready to use. The hinge angle offset estimation and error propagation together with the measurement statistics in Chapter 3.3 were all implemented in MATLAB. As input to the hinge angle offset estimation method a real log-file was used. A script was also made to be able to go through a number of log-files to analyse the confidence interval of the method using the measurement statistics in Chapter 3.3.4. All laser offset estimation parts in Chapter 3.4 were implemented in C++ using Microsoft Visual Studio 2005 and used data in a log-file together with a configuration file with all parameters as input. The laser offset implementation was built on modifications of an already existing Route Profiling and Playback code. From the C++ programs mat-files were delivered that could be analysed by MATLAB. A script was written in MATLAB to be able to automatically run estimations from many log-files automatically. 4 Error propagation in the hinge angle method To see how the errors propagated through the hinge angle offset estimation two analyses were made using the error propagation in Chapter 3.3.3. The analyses ˆ& shows how the unbiased angular speed measurement q (n), speed ubm measurement v (n) and hinge angle measurement g (n) affects the standard m m deviation of the estimated offset sˆ . A third analysis was made to see the gˆ o effect of just iterating the hinge angle estimation ones using g (n) as the m estimated hinge angle gˆ(n) in Eq. ( 7 ). 4.1 Analysis setup All analyses were made at the systems sample frequency f of 25 Hz. An early s analysis shown that the error isn’t affected by g& (n) when smaller than 14 o/s, m which is maximum steering speed, so no further analysis for it was made. In all 21
Chalmers University of Technology
experiments g (n) was swept from 0 to 45o, where 45o is the maximum steer m &ˆ angle. q (n) was calculated from Eq. ( 1 ) using v (n) and g (n) as inputs. ubm m m &ˆ q (n), v (n) and g (n) were then inserted to the error propagation equations ubm m m to give s . gˆ o Different analyses were made starting with investigating how the standard deviation behaved when only using one sample at different speeds v (n) equal m to {1; 2.5; 4.5} m/s. Using more samples N only decrease sˆ with one divided gˆ o by the square root of N and doesn’t change the characteristic. When travelling at low speed more samples N are collected compared to travelling at high speed. It was investigated how this affected sˆ using Eq. ( 11 gˆ o ) at the speeds v (n) equal to {1; 2.5; 4.5} m/s when travelling 50 m. m A final analysis was made to evaluate the error in gˆ from using g (n), o m including the offset g , as the estimated hinge angle gˆ(n) in the denominator of o Eq. ( 7 ) and iterate the hinge angle estimation method only one time. It should be noticed that the error is a constant error that depends on g and not a random o error as those above. Here g& (n) affects the error and according to Eq. ( 7 ) does larger g& (n) give larger effects, so g& (n) was put to its absolute maximum value of 14 o/s. The error also increase with decreased speed so to find the maximum error the speed was put to the lowest speed used around 1.0 m/s. Increased offset also increases the error according to Eq. ( 7 ). To see the effect of different offsets g , the set {2, 5, 25, 45}o was tested. o 4.2 Results Using only one sample to estimate the offset gives the standard deviation sˆ gˆ o seen in Figure 4 at different hinge angles g (n) and speeds v (n). As seen sˆ m m gˆ o increases with g (n) and decreases with increased speed v (n). m m The result for travelling 50 m, giving different number of samples, can be seen in Figure 5. As seen in the figure sˆ increase less for large angles and low gˆ o speeds compared with Figure 4. Analyses of the effects of using g (n) as gˆ(n) in Eq. ( 7 ) showed that the error m in gˆ increased almost linearly with increased g (n) as seen in Figure 6. It o m could also be seen that the error in gˆ is always smaller then the true offset. o 22
Chalmers University of Technology
Figure 6: Error in estimated offset at the known offsets {2, 5, 25, 45}o because of using g (n) as gˆ(n) in Eq. ( 7 ). It should be observed that the error is m always less than the true offset. 4.3 Discussion From Figure 4 and Figure 5 it can be concluded that the estimated standard deviation sˆ increases with increased hinge angle g (n) and decreases with gˆ m o increased speed v (n). The reason for increased sˆ with increased g (n) is m gˆ m o because of that uncertainty in speed increases the uncertainty in rotation speed & q with increased g (n). If the uncertainty in speed is zero then sˆ is almost m gˆ o & constant for increased g (n). Increasing v (n) gives increased q when keeping m m g (n) constant and thereby better readings. It should also be remembered that m according to Eq. ( 11 ) increasing the sample size N by driving a longer distance decreases sˆ . From Figure 5 it could be concluded that when driving 50 m the gˆ o 99 % confidence interval 2.576s is only 0.21o even for low speeds at small gˆ o g (n), which is close to ±0.2o. m When analysing the error introduced because of using g (n) as gˆ(n) in Eq. ( 7 m ) it was found that the error increased almost linearly with g (n). It was also m found that the error in gˆ was always smaller than the true offset which makes o it possible to iterate to find the best estimate gˆ of the true hinge angle. The o error is a constant error, which means that taking more samples doesn’t help, but iterating as described in Chapter 3.3.1 reduces the error to zero. 24
Chalmers University of Technology
A conclusion is that because of increased sˆ at high hinge angles it’s expected gˆ o that the machine is going to be driven as straight as possible. When using the mean of g (n) as an first approximation of the offset this gives a remaining m offset in the hinge angle estimate gˆ(n) equal to the mean of the hinge angleg (n), see Eq. ( 8 ). If g (n) is small the error is even smaller according to Figure 6 so as long as the machine is not driven with to high g (n) no iteration is needed. 5 Hinge angle experiments To test the calibration method a series of experiments were needed. A list of goals below was created to fully test the systems limits. • Find the minimum length needed to get good enough precision. • Check the precision in relation to speed. Find minimum speed. • Check the precision in relation to turn radius. Find maximum angle. • Check the deadband of the machine. • See if stretching the machine in one direction gives better performance. 45 experiments to achieve these goals were put up. In most of them one drove straight or in an arc. No experiment was made with a large hinge angle offset because the theoretical analysis above proves that it doesn’t work. The experiments were made on an ST14 LHD in an old underground mine in Kvarntorp in Kumla by an experienced driver. The data were then saved analysed offline in MATLAB. The third goal couldn’t be tested because of limited space in the mine drift. 5.1 Experiment setup All experiments were repeated at the speeds {1.2; 2.5; 4.5} m/s because they are the maximum speeds at each gear and thereby easy to keep constant. In all cases the driver tried to drive relaxed so normal correction oscillations should be included. The hinge angle sensor was pre-calibrated to give a reasonable hinge angle. Then the mean of the sampled hinge angles when driving straight 300 m was calculated, which gave the remaining true hinge angle offset with a precision of around ±0.05o. The true offset was then used as a reference in the experiments. To get the gyro bias the machine stood still for 10 s before and after completing a run through the drift to collect gyro bias data. The first experiments were made to test the method under normal conditions. To collect data the machine was driven straight one to two times back and forth through a 300 m long tunnel at the different speeds. A straight path was chosen because that should give the smallest errors. To test the last two goals the machine was driven with the true hinge angles of {0.2; ±0.7}o at different speeds a distance of 100 m in an 11 m wide tunnel. The 25
Chalmers University of Technology
hinge angle of 0.2o was chosen to see if better performance was achieved if the machine was stretched in one direction. The thought was that if one stretches the machine in one direction the machine should not oscillate because of the play in the axles and the hinge. This then should increase the performance of the method. ±0.7o were chosen to see if there was any play in the hinge angle sensing and thereby different results in the offset estimation. The reason to not test with random moves was that the errors arise when the angle is high and/or speed low. Also the error that comes from increasing/decreasing the angle is small. If totally random moves had been made the errors had cancelled out. Enough with data were collected at each setup to be able to divide it into a number of partions containing N measurements where N was chosen to 250, 2 , 2 500 and 1000 samples, to test the precision at different lengths. Each experiment setup was repeated N times to find the confidence interval of the 1 estimated offsets. A number of statistics were then calculated using the data starting with the mean gˆ of all N offset estimatesgˆ to see that they were centred on the true o 1 o offset. Then the maximum deviation from the true offset d was calculated gˆ max o for each experiment setup. d is the maximum error from our gˆ max o measurements, which is only a small amount of the whole population. Confidence intervals take the whole population into consideration and are more assertive statistics. The methods described in Chapter 3.3.3 and 3.3.4 were therefore used to find different 99 % confidence intervals for gˆ . For each o measurement n in experiment p the 99 % confidence interval was calculated using the error propagation with the measurements as input. The largest confidence interval z max(sˆ (p)), where z =2.576, was then recorded 0.99 gˆ 0.99 o for each setup. The largest and smallest confidence interval z max(s (p)) respectively 0.99 gˆ 1 o z min(s (p)), using the standard deviation in each experiment p, were 0.99 gˆ 1 o recorded for each setup. By using the estimated standard deviation s of all the gˆ o gˆ :s, the confidence interval with lower boarder z L and higher boarder o 0.99 1 z L was calculated and recorded for each setup together with the estimated 0.99 2 c 2 confidence interval z s and the mean of all N ratios for each 0.99 gˆ o 1 N - 1 1 experiment setup. 26
Chalmers University of Technology
5.2 Results From the mean value, when driving 300 m straight, it was found that the true remaining hinge angle was 0.31 o ±0.05 o. So the goal for the experiments is to get hinge angle offsets in the range 0.31 o ±0.20 o, to show that the method can operate under all conditions. 5.2.1 Driving straight Table 1: (a) The experiment setups, mean hinge offsets gˆ and maximum o deviating offsets d when driving with a true hinge angle g of 0o. (b) The gˆ max o Gaussian 99 % confidence intervals for the experiments using different methods. a. g Exp. v Setup [o] Gear [m/s] N 2 f s Distance N 1 gˆ o d gˆ omax 1 0 1 1.3 250 20 16.25 19 0.34 0.36 2 0 2 2.6 250 20 32.50 38 0.30 0.20 3 0 3 4.5 250 20 56.25 38 0.32 0.15 4 0 1 1.3 500 20 32.50 9 0.34 0.20 5 0 2 2.6 500 20 65.00 18 0.30 0.11 6 0 3 4.5 500 20 112.50 18 0.32 0.11 7 0 1 1.3 1000 20 65.00 4 0.34 0.19 8 0 2 2.6 1000 20 130.00 6 0.30 0.05 9 0 3 4.5 1000 20 225.00 6 0.31 0.05 b. E Sex tp u. p z 0.99max(sˆgˆ o(p)) z 0.99min(s gˆ o1(p)) z 0.99max(s gˆ o1(p)) z 0.99L 1 z 0.99sgˆ o z 0.99L 2 Nc 12 - 1(p) 1 0.26 0.08 0.20 0.20 0.29 0.49 6.40 2 0.13 0.06 0.19 0.14 0.19 0.26 5.39 3 0.08 0.04 0.19 0.12 0.15 0.21 7.94 4 0.19 0.07 0.11 0.15 0.25 0.60 9.47 5 0.10 0.04 0.13 0.10 0.14 0.23 4.06 6 0.06 0.03 0.12 0.08 0.11 0.19 9.88 7 0.13 0.05 0.07 0.14 0.28 1.79 23.23 8 0.07 0.03 0.06 0.03 0.06 0.24 2.61 9 0.04 0.03 0.06 0.04 0.07 0.25 7.30 The results from driving straight with different speeds and using different number of samples can be seen in Table 1, for figures see appendix B. As seen, the estimated offsets gˆ have a mean gˆ very close to 0.31o in all cases. If one o o looks at their distributions one finds out that they have Gaussian distributions which tell us that our assumptions made in Chapter 3.3.3 and 3.3.4 are valid. From Table 1 it could be seen that for all cases, besides the one when only using 250 samples and driving with the first gear, d is smaller than or gˆ max o equal to 0.20o. 27
Chalmers University of Technology
Using gear 2 and 3 instead of gear 1 largely improves the performance according to d . Also increasing the distance driven and thereby the sample gˆ max o size decreases d for gear 2 and 3 but only a small amount for gear 1 at gˆ max o larger sample sizes. The difference between using gear 2 and 3 decreases with increased sample size. Table 1b shows the positive borders of different 99 % confidence intervals calculated using the methods in Chapter 3.3.3 and 3.3.4. z L and z L tell 0.99 1 0.99 2 in which interval the true positive confidence interval border is. They only relies on that the measurements are Gaussian and are therefore the most reliable ones. z max(sˆ (p)) is the most uncertain statistic. As seen in Table 1b it 0.99 gˆ o compared to z L underestimates the confidence interval in most cases. 0.99 1 z s (p) also underestimates the confidence interval sometimes because of 0.99 gˆ 1 o that z min(s (p)) is smaller than z L for all setups and sometimes leave 0.99 gˆ 1 0.99 1 o reasonable estimates z max(s (p)), which in most cases are in the interval 0.99 gˆ 1 o defined by z L and z L . But on average it underestimate the confidence 0.99 1 0.99 2 c 2 interval according to the mean of . N - 1 1 It can be seen from z s that using confidence intervals instead of d 0.99 gˆ gˆ max o o gives similar results. By looking at z L it can be seen that for gear 2 and 0.99 2 gear 3 the error is smaller than 0.3o in all cases if one includes the measurement error of the true offset. Increasing the distance travelled from 33 m doesn’t improve z L . To see the spread of the offsets and their confidence intervals 0.99 2 see appendix B. 5.2.2 Small turning Table 2 (a) The experiment setups, mean hinge offsets gˆ and maximum o deviating offsets d when driving with a true hinge angle g of 0.2o. (b) The gˆ max o Gaussian 99 % confidence intervals for the experiments using different methods a. g Exp. Setup [o] Gear v [m/s] N 2 f s Distance N 1 gˆ o d gˆ omax 10 0.2 1 1.3 250 25 13.00 12 0.30 0.25 11 0.2 2 2.6 250 25 26.00 12 0.27 0.18 12 0.2 3 4.5 250 25 45.00 12 0.37 0.25 13 0.2 1 1.3 500 25 26.00 4 0.29 0.13 14 0.2 2 2.6 500 25 52.00 4 0.28 0.05 15 0.2 3 4.5 500 25 90.00 4 0.39 0.18 28
Chalmers University of Technology
By looking at the gˆ at ±0.7o in Table 3 it’s found that turning positive o increases the estimated hinge angle gˆ and turning negative decreases gˆ . It o o could also be seen that the deviation from the true hinge angle offset decreases with speed. If Table 2 and Table 3 are compared it can also be observed that increased turning also increases the deviation in gˆ . One exception is at gear 3 o in Table 2 where the deviation is equally large as those at other gears in Table 3. To see the spread of the offsets and their confidence intervals see appendix B. 5.3 Discussion From the results when driving straight one can draw the following conclusions: 1. Performance improves with higher speed and longer distance. 2. The difference of using gear 3 and 2 decreases with distances travelled. 3. Driving slower than 2.5 m/s and shorter than 33 m is not recommended. As seen in Table 1 both z sˆ (p) and z s (p) are underestimating the 0.99 gˆ 0.99 gˆ 1 o o true confidence interval. One of the reasons for z sˆ (p) to underestimate 0.99 gˆ o the confidence interval is that it uses a model, which in reality always contains model errors. It also only takes into consideration the standard deviations of the sensors that for the speed and the gyro contain modelling errors. Errors like slippage and vibrations, which affects the gyro, is also not considered. Even when z sˆ (p) underestimates the confidence interval it still models the 0.99 gˆ o behaviour correctly, when compared to z s in Tables 1, 2 and 3. 0.99 gˆ o Driving at 4.5 m/s didn’t give any significant improvement in both the straight case and when a small turn was made. It gave the same confidence interval, driving 56 m at 4.5 m/s, as driving 65 m at 2.5 m/s, as seen in Table 1. Driving slower makes it easier to perform the calibration in small and narrow mines so 2.5 m/s is preferred to 4.5 m/s. z s (p) only rely on the facts that the measurements are not correlated, that 0.99 gˆ 1 o the sample size is large and that they have a Gaussian distribution. The later has been confirmed and the sample size is large. By calculating the correlation it was found that there exists an oscillation with a period of around 1 s in the sample offsets. By looking at the gyro readings and the machine when the machine stops it has been seen that the LHD is swaying for some seconds with a period of 1 s. From this it can be concluded that the LHD has a swaying resonance frequency of around 1 s and because of the correlation the variance is under estimated. The error introduced by the swaying is small if large sample sizes are used. The experiments made when turning with a small hinge angle shows that there is nothing gained in precision by doing it. It’s even so that the precision seems to decrease according to Table 2. When using 500 samples and gear 2, Table 2, 30
Chalmers University of Technology
it seems like that the confidence interval is largely decreased, but looking at z L one understands that it is possible that it’s underestimated. These results 0.99 2 are also backed up by the mean ratio, which normally is larger than 1. The reason for the mean offset estimate gˆ to be high at high speeds can’t be o explained by other than the occurrence of slippages at high speed when turning. This is probably also the reason for the increased confidence interval, but it should be remembered that there exists a uncertainty in gˆ . o Another conclusion is that when the turning is increased a deadband is created according to Table 3. It’s not a normal mechanical deadband because it shouldn’t increase with turning angle. More likely it appears because of slippages. It can also be observed from z L and z s in Table 1, 2 and 3 0.99 1 0.99 gˆ o that the confidence interval and thereby the spread increases with increased hinge angle. From this a conclusion can be drawn that driving as straight as possible is recommended. It should be noted that the z L to z L intervals are too large and 0.99 1 0.99 2 overlapping, because of too few experiments with each setup, in order to with 99 % confidence assure the above conclusions. But there is a high probability that they are right. It should be remembered that z L and z L depends on 0.99 1 0.99 2 the number of samples and therefore converges to the true confidence interval when the number of experiments with the same setup goes to infinity and vice versa. Without more data it can only be assured that the offset estimation error is less than z L . A effect of this can be seen in the straight case where it 0.99 2 seems as increasing the distance doesn’t decrease z L . This happens because 0.99 2 the sample size decreases with distance. If one instead looks at the estimated confidence interval z s it decreases. The problem of too few samples is also 0.99 gˆ o one of the factors that drives z L high in the cases when making small turns. 0.99 2 From the above experiments it can be concluded that the method can deliver a hinge angle offset with a precision better than 0.3o with a 99 % confidence as long as the machine is driven straight 33 m or longer at a speed of 2.5 m/s. The goal of 0.2o can’t be achieved following the criterions in Chapter 2.1 but it is possible, with high probability, that more experiments is going to show that the true 99 % confidence interval lies lower than the here estimated L :s. 2 Later investigations of the LHD used at data collection in Kvarntorp showed that a part in the hinge angle sensor was broken. This gave the hinge angle sensor a play. The load on the hinge angle changes when the machine goes in different directions which then probably affected its readings because of the play. This can have altered the results negatively and is probably the reason for the deadband. 31
Chalmers University of Technology
6 Required laser offset accuracy For the lasers the required accuracy was specified as a maximum scan matching error in the map. To be able to specify the corresponding required precision of the angular offset of the lasers a set of offsets in a number of environments were tested to find the ones giving the maximum allowed scan matching error of 0.35. 6.1 Analysis setup To find the offsets giving the maximum allowed scan matching error a grid of varying laser offset combinations (f ,f ) was created. It was expected that the R F laser offsets never would be larger than ±3o degrees. But we chose to use a grid with the range [-5 5]o in each direction and a resolution of 1o. The grid was applied on real log data from six different routes in a mine in the Finnish town Kemi. Three routes were straight and three routes had a turn. For each grid point and route the mapping and evaluation procedure described in Chapter 3.4.1 was applied using the offsets in the grid as correction. The reference offsets were estimated from the three straight routes using the above method with a resolution of 0.3o and were found to be equal to (-0.8;0.3)o, which then were represented by the vector φ . ref 6.2 Results Figure 7: Highest scan matching error from all routes at each offset combination. Green crosses and the red shaded area in the middle show the offsets giving a e smaller than 0.35. The red circle shows the offset φ match ref and the red square shows the offsets φ with lowest e . Goptima match Figure 7 shows the total maximum of the scan matching errors, when driving the paths with different offsets. The green crosses and the red shaded area in the middle show the offset combinations that give a e smaller than 0.35 for all match 32
Chalmers University of Technology
In Figure 8a and 8b the maximum scan matching errors from straight paths respectively paths with a turn can be observed. Figure 8a from the straight paths shows that the offsets φ with loweste in the straight cases coincides SOptima match with φ . It also shows that the area of allowed offset is larger than in Figure ref 7, centred around φ and allows the offset to vary ±2o. These results are also ref valid if one looks at the three paths separately. It could be seen that Figure 8b from the turning paths resembles Figure 7 since it has a higher e for almost match all angles and thereby dominates Figure 7. It should however be mentioned that it’s only one path that dominates the other three and decides the shape of Figure 7 and 8b. In all turning cases φ varied but the area of allowed offsets was TOptima centred around φ and had the same size as in the straight cases with ref exception of the dominating one. 6.3 Discussion It’s the worst cases that set the limits because the scan match error must always stay under 0.35. From Figure 7 the conclusion can be drawn that the offsets can’t be allowed to deviate more than -0.5o from their true value. One also assumes that one can’t allow them to deviate more than +0.5o. This conclusion comes from the symmetry in the systems dimensions which should give symmetric scan matching errors and is confirmed by the straight cases where the allowed area is centred around φ . More cases had been needed to true investigate this further but because of too large computation times as well as lack of relevant data this was not possible. Figure 9: Data from one of the turning paths. As seen there exist an offset that gives a lower maximum scan matching error at the turn than the true offset. 34
Chalmers University of Technology
The reason for φ to sometimes deviate from φ in the turning cases can TOptima ref be explained by Figure 9 showing data from one of the turning paths. In Figure 9 the blue line is e using φ , the red line is e using φ and the match ref match TOptima green line is the hinge angle g in radians. From the figures it can be observed that e goes high when g goes high. φ gives a better e on average match ref match but φ gives the lowest maximum e . One of the reasons for e to TOptima match match increase in turns is because of the geometry of the mine drift and the side drift, which the turn is made into, because e is also increasing when just passing match a side drift, which can be seen by looking at e for each sample in the match straight cases. Sometimes there are also walls with ventilation tubes or cable ladders that make the laser data corrupt and increasee . e also goes down at erroneous match match offsets because of above mentioned reasons. As a result of that we didn’t collect the data ourselves, we can’t know where there were things interfering with our measurement. Nevertheless, non true offsets have shown to sometimes give better scan matching errors. 7 Accuracy of the laser offset estimation To evaluate that the method provide reliable offsets with high enough precision a set of experiments were performed on real and simulated data. 7.1 Test setup Table 4: Places, angles used at simulation and number of real tests at each location. Simulated (f ;f ) [o] Real [number of tests] R F Kemi Straight (0,0), (0,3), (3,3) 10 Kemi Turn (0,0), (0,3), (3,3) 3 Kvarntorp (0,0), (0,3), (3,3) 5 A test schema seen in Table 4 was created with different conditions to test. It contains tests to see if the method could handle both straight and turning paths in real and simulated cases. The tests were done using log-files from a mine in Kemi, which has representable proportions. The simulation tests were made to see that the method could handle all expected offsets f and f because we R F couldn’t do real experiments where we turned the lasers. Routes from the same area in Kvarntorp were used to test how it handled large mine drifts with many side drifts. It was also of interest to see how fast it converged in the different cases and how long driving distance that was needed. The method shouldn’t need reconfiguration for each environment so in all tests the same parameter configuration was used. A speed of around 2.5 m/s was used because the early tests had shown that higher speeds gave the best result and 2.5 m/s is the maximum usable speed. The hinge angles measurements of 35
Chalmers University of Technology
the real data where compensated for sensor offset using the method described in Chapter 3.3.1. In the simulations the same driving path was used in each location for all offsets to not affect the results. 4o turns were made in the end of the Kvarntorp runs to see if the Kalman filter could handle it. When analysing the real paths the true offsets were known for the Kemi data from the analysis above and equal to (-0.8;0.3)o. In Kvarntorp the true offsets weren’t known but constant in all routes. The reason to use real cases was to be able to analyse the accuracy under real circumstances. In the straight Kemi cases 10 different log-files had been recorded in parts of two areas in the mine. As a result of this the following tests are in the same areas • 1, 3, 4 and 8 • 6, 9 and 2 • 7, 10 and beginning of 5 The Kalman filter uses the scan matching error e to update its estimates match and a larger e makes the algorithm to converge faster. As seen in Figure 7 match equal offsets and opposite offsets give almost the same scan matching error e . If (f ,f ) are (3,0)o or (0,3)o it should be harder than if they are (3,3)o to match R F find them, due to lower scan e . But the convergence shouldn’t depend on if match the offsets are (3,0)o or (0,3)o. From this a conclusion was drawn that it’s enough to test with the offsets (f ,f ) equal to {(0,0) (0,3) (3,3)}o. R F 7.2 Results 7.2.1 Driving straight in a normal mine drift f f R F Figure 10: f and f from the simulated straight path in Kemi at laser offsets R F {(0,0) (0,3) (3,3)}o. 36
Chalmers University of Technology
In Figure 11 f , f , fˆ and fˆ can be seen from data recorded along real R F R F straight paths in Kemi, see maps in Figure 12. From looking at the figure it can be seen that the offsets are spread around (f ,f ) equal to (-0.8; 0.3)o, which is R F the true offset. 8 of the 10 final (f ,f ):s are inside the ±0.5o limit and the R F same 8 are inside the limits after only 23 m. By looking at (fˆ ,fˆ ) it can be R F seen that some estimates converge fast and some takes longer time, especially for fˆ . All of them have converged close to their final value after 40 m and F (f ,f ) have get close to their final values after around 50 m. The computation R F time was around 17 minutes in all cases. 7.2.2 Turning in a normal mine drift f f R F fˆ fˆ R F Figure 12: Angular offsets estimated from a simulated route in Kemi with a turn and laser offsets {(0,0) (0,3) (3,3)}o. As seen, fˆ get affected at the turn F after 55 meters. Figure 12 shows f , f , fˆ and fˆ when simulating with the offsets {(0,0) R F R F (0,3) (3,3)}o. The offsets have converged to within ±0.1o when the mean valuing is started after 23 m and doesn’t get largely affected by the turn after 50 m. As seen by (fˆ ,fˆ ) the offsets converge after a few meters. fˆ does get R F F affect in the turn but then returns to its original value. For fˆ it’s hard to see if R it is affected by the turn. Its largest reaction happens after 70 m, which is when the back has left the turn and is facing towards it. 38
Chalmers University of Technology
f f R F fˆ fˆ R F Figure 13: Angular offsets estimated from real data recorded along a path including a turn. The same accuracy was not achieved when using real data in real turning cases. This can be seen in Figure 13 showing f , f , fˆ and fˆ when driving the R F R F routes seen in Figure 14, starting at (0,0) m with heading 0o. Test 3 is the only case when both offsets are within the ±0.5o limit and close to the true offsets. Test 2 is on the border of the ±0.5o range with f and close but outside with F f . In test 1 only f is in the range but by looking at fˆ in Figure 13, it’s seen R R F that fˆ after the turn, seen in Figure 14, converges towards the true offset. It F could also be seen that fˆ in test 3 deviates at the turn after 60 m and then F returns. In test 2 the offset is deviating all the time from the true but especially when the turn is made. The computation time is around 15 minutes. 39
Chalmers University of Technology
7.2.3 Driving in a straight wide mine drift Simulations from Kvarntorp gave offsets (f ,f ) within ±0.2o of the true R F offsets. The simulations also showed that the effects of the small turns in the end are neglect able. Figure 15 shows f , f , fˆ and fˆ estimated from data recorded in R F R F Kvarntorp. For all the Kvarntorp runs the offsets are the same but not known and it’s hard by just looking at Figure 15 to see what the true offsets could be because of the large spread. The worst ones are the rear offsets which are spread ±0.25o at all times. Even though it’s hard to see the true offsets it can be seen in Figure 15 that (f ,f ) are inside the ±0.5o limit after 23 meters if the R F true offsets are somewhere between the estimated ones. The computation time is around 20 minutes. From looking at (fˆ ,fˆ ) in Figure 15 it could be seen that for both offsets test R F 1, 3 and 5 are grouped together and similar with test 2 and 4. Figure 16 shows the driven paths super imposed on maps of the environment. All paths begin in position (0,0) m with heading 0o. In tests 1, 3 and 5 the machine was driving forward, while 2 and 4 are recorded while driving backwards. It’s thereby a correlation between the direction and especially the rear offsets. From Figure 15 it can be seen that it takes 50 m for (fˆ ,fˆ ) to stabilize and because the R F paths are not longer than 80 m (f ,f ) do not have time to stabilize at the final R F value. The noise on the maps in Figure 16 is because of holes in the wall into a side drift. Figure 16: The 5 maps and paths, in blue, from Kvarntorp. The paths begins at position (0,0) m with heading 0o. 41
Chalmers University of Technology
7.3 Discussion In Chapter 6 it was found that an accuracy of ±0.5o was needed to get a scan matching error less than 0.35. It’s here seen that it’s possible in most cases to achieve this accuracy after only 23 m if the machine is going straight. Possibly it could have been achieved faster if the mean valuing had started earlier since the offsets (fˆ ,fˆ ), according to Figure 11, in most cases have converged after R F only a few meters. The computation time is also less than the required 30 minutes. The reason for tests 3 and 4 in Figure 11 to lie outside the allowed range can’t be properly explained. Tests 1 and 8 are made in the same drift and in test 1 there are only minor problems to converge in the beginning. In test 8 there are larger problems in the beginning and the start position is in the same place as in tests 3 and 4. It may be that the machine is starting in a junction, resulting in the estimated offsets getting stuck in local minima. It could also be observed that the wall is noisy after half the path. This may also be the reason why tests 3 and 4 don’t converge after the junction. Investigations of e have shown that it decreases for tests 3 and 4 compared to in test 8 at match the place of the noisy wall, which strengthens the theory. This happens because the noisy wall changes the optimal point searched for by the Kalman filter. The noise may come from a cable ladder or ventilation drum. What is causing the constant offsets between the tests in the same areas has yet to be proven. Probably it has to do with differences in the starting point, steering of the machine, noise in the gyro and so on. A probable reason for (fˆ ,fˆ ) to deviate from their true offsets in the R F simulated and real turns in Figure 12 and Figure 13 is because of geometry of the turn, as mentioned in Chapter 6.3. The theory is strengthen by the fact that fˆ in one of the simulated cases isn’t affected until after the turn, and that we R don’t have simulated odometry errors. Test 2:s behaviour, seen in Figure 13, can’t be explained. Something makes the offset deviate even on the straight parts. Later examination of the machine used at data collection in Kvarntorp showed that a part in the hinge angle sensor was broken. This gave the hinge angle sensor a play. The load on the hinge angle changes when the machine goes in different directions, which then probably affected its readings because of the play. This is probably the reason for the grouping in the tests seen in Figure 15. Positioning of the rear part of the machine depends directly on the hinge angle, which can be the reason for the rear offsets to get more affected. The algorithm works in all simulated cases which mean that it is theoretically correct. More experiments in a controlled environment are needed to draw some definitive conclusions for the real cases but from the experiments made it can be concluded that the offsets in most cases can be estimated to within ±0.5o, if the environment doesn’t contain too much disturbance. The requirements are 42
Chalmers University of Technology
that the machine is driven straight for at least 23 m at a speed of around 2.5 m/s. It even seems to work in wide drifts with side drifts. The computation time increases with the width of the drift but stays below 30 minutes. 8 Conclusion This report presents two methods that have been developed for use in calibration of the hinge angle sensor respectively the two laser scanners on an Atlas Copco ST14 LHD. In both the hinge angle calibration and laser scanner calibration the new methods finds the angular offset of the sensors. For hinge angle calibration we have developed an algorithm based on the difference between the measured hinge angle value and the expected value modelled from the speed and the heading change as measured by the gyro. In the laser scanner calibration we used state augmentation together with an already existing Kalman Filter to estimate the angular offsets of the laser. A SLAM algorithm was also implemented to solve the problem with cross dependence between the map and the offsets at the calibration. Both methods satisfy the predefined criterions that no surrounding equipment should be needed and no extra sensors should be needed. They also satisfy the requirement of easy operation. The operator only needs to drive approximately straight for around 50 meters while recording data and then supply the calibration program with the log-file. Both methods can also operate offline and deliver estimates in under 30 minutes as requested. To test accuracy and dependence on outer parameters a number of experiments and tests were made with both methods. The dependence on speed, distance and steering angle was analytically investigated for the hinge angle calibration to get the characteristics of the system. Real experiments were then made to confirm the characteristics and find the limits of the system. For the laser scanners the required precision was only known as a maximum scan matching error in the map and not as a sensor offset. An investigation was therefore first made to find the offsets that gave the maximum scan matching error. The system was to complex to do analytical investigations, instead simulations and real tests were made to test the accuracy of the system and how long driving distance and time that was needed to do the estimation. The tests were made in normally sized tunnels, tunnels with a turn and wide tunnels with side drifts to test how the method handle different environments. The experiments showed that the hinge angle calibration method is only able to deliver accurate estimates two within ±0.3o in 33 m or more. It is also required that the machine drive straight for this to be achievable. The tests showed that the laser calibration method is able to deliver the required accuracy of ±0.5o in around 15 minutes, and that a distance of 23 m in most cases is sufficient if the machine is driven straight. Problems arise if there are to much environmental disturbance or if the machine makes large turns. For both methods a speed of 2.5 m/s is recommended. 43
Chalmers University of Technology
The laser calibration method and the hinge angle method both need further evaluation in the future. More experiments needs to be done to better establish the confidence interval of both methods. As explained above side drifts interfere the laser offsets estimation and maybe also the positioning so it needs to be investigated further. In Atlas Copco auto traming system maps are today created using only odometry. In the future one could use the maps from the laser calibration method instead where the positioning is made using both odometry and laser measurements. One could also augment the hinge angle sensor offset to the states and use the laser calibration method for the hinge angle offset estimation. The hinge angle method could also be run online, when the offset is known to keep track of the gyro and hinge angle sensor to see if an error occur, because then the method deliver a different value compared to the true offset. Finally the developed methods considerably simplify the calibration of the three sensors compared to the methods used before. Both methods delivered such a good results that Atlas Copco choose to implement them in their system. 9 References Altafini, C. (1999). A path-tracking criterion for and LHD articulated vechicle. International Journal of Robotics Research, Vol. 18, No. 5, pp. 435-441. May 1999. AutoMine (2011). Available: http://www.miningandconstruction.sandvik.com. [2011-02-02] Borenstein, J. (1996). Measurements and Correction of Systematic Odometry Errors in Mobile Robots. IEEE Transactions on robotics and automation. vol. 12, No. 6, pp. 869-880. December 1996. Dall Larsen, T., Bak, M., A. Andersen, N. & Ravn O. (1998). Location Estimation for and Autonomously Guided Vehicle using and Augmented Kalman Filter to Autocalibrate the Odometry. IEEE International Conference FUSION. (pp. 245-250). Las Vegas, Nevada, USA 6-9 July 1998. De Ceccco, M. (2002). Self-Calibration of AGV Inertial-Odometric Navigation Using Absolute-Reference Measurements. IEEE Instrumentation and Measurement Technology Conference. (pp. 1513-1518). Anchorage, Alaska, USA 21-23 May 2002. Hainsworth, D.W (2001). Teleoperation User Interfaces for Mining Robotics. Autonomous robots, vol. 11, ss. 19-28. Larsson, J., Broxvall, M. & Saffiotti, A. (2010). An evaluation of local autonomy applied to teleoperated vehicles in underground mines. IEEE International Conference on Robotics and Automation (ICRA) (pp. 1745- 1752). Anchorage, Alaska, USA 3-8 May 2010. Madhavan, R., Dissanayake, M.W.M.G. & Durrant-Whytel, H.F. (1998). Autonomous Underground Navigation of an LHD using a Combined ICP-EKF 44
Chalmers University of Technology
Nitrogen removal in process water from the Gerum tunnel A lab analysis using nitrification and denitrification as biological treatment Master of Science Thesis in the Master’s Programme Infrastructure and Environmental Engineering VICTORIA LILJEDAHL Department of Civil and Environmental Engineering Division of Water Environment Technology Chalmers University of Technology ABSTRACT High nitrogen concentrations in process water from tunnel constructions have become an issue due to residues from explosive, transported to recipients in the construction area. Biological treatment in terms of nitrification and denitrification are necessary to remove the nitrogen compounds in the process water. Skanska Sweden AB is currently constructing a tunnel in Tanumshede called Gerumstunneln, for which the biological treatment of the process water has not worked during the project why this study was initiated. The aim of this master thesis was to assess which factors and parameters affect the biological treatment of process water in tunnel construction with regard to nitrification and denitrification. Lab tests on process water from the tunnel project were performed to achieve a complete nitrification and denitrification. Three nitrification tests and one denitrification test were performed in the lab where the first nitrification test used process water from two separate days and sludge that was collected from a wastewater treatment plant. The nitrification process was incomplete since nitrate did not increase even though there was a slight decrease in ammonium and a slight increase in nitrite. The second test comprised of the same water and sludge materials as Test 1 with the exception of adding sodium dihydrogen phosphate, thereby increasing the phosphate concentrations which were thought to limit the reactions. This test also failed to show any nitrifying activity. The alkalinity measured in the process water was very low and for the third test, alkalinity and phosphate were therefore increased. This test gave the best results of the nitrification tests and resulted in ammonium being converted to nitrite, however the process failed to transform nitrite to nitrate. The results from the denitrification test did show a denitrifying activity when nitrate was transformed to nitrite at the same time as total nitrogen (TN) decreased which indicates that nitrogen gas has been formed, though the test should have been carried out for a longer period of time. Further lab analysis on process water in tunnels with high nitrogen content is crucial which should focus on metal concentration and the proportions of sludge and process water since those parameters were not thoroughly analyzed in this study. Key words: nitrogen, nitrification, denitrification, process water, tunnels, inhibiting factors, water assessment II
Chalmers University of Technology
Kväverening av processvatten från Gerumstunneln En labbanalys där nitrifikation och denitrifikation används som biologisk rening Examensarbete inom Infrastructure and Environmental Engineering VICTORIA LILJEDAHL Institutionen för bygg- och miljöteknik Avdelningen för Vatten Miljö Teknik Chalmers tekniska högskola SAMMANFATTNING Höga kvävehalter i processvatten från tunnelkonstruktion har blivit ett problem på grund av rester från sprängmedel som transporteras till vattenrecipienter inom byggområdet. Biologisk behandling såsom nitrifikation och denitrifikation är nödvändigt för att avlägsna kväveföreningar i processvattnet. Skanska Sverige AB bygger för närvarande en tunnel i Tanumshede, Gerumstunneln, där den biologiska reningen av processvattnet inte har fungerat under projektet varför denna studie påbörjades. Syftet med detta examensarbete var att bedöma vilka faktorer och parametrar som påverkar den biologiska reningen av processvatten under tunnelkonstruktion med avseende på nitrifikation och denitrifikation. Laboratorietester med processvatten från tunnelprojektet utfördes för att uppnå fullständig nitrifikation och denitrifikation. Tre nitrifikationstester och ett denitrifikationstest utfördes i labb där det första nitrifikationstestet använde processvatten från två olika dagar och slam som samlats in från ett avloppsreningsverk. Nitrifikationsprocessen var ofullständig då nitrat inte ökade trots att det fanns en liten minskning av ammonium och en liten ökning av nitrit. Det andra testet bestod av samma vatten - och slammaterial som Test 1 med undantaget att natriumdivätefosfat tillsattes, vilket ökade fosfatkoncentrationerna so ansågs begränsa reaktionerna. Detta test misslyckades också med att visa någon nitrifierande aktivitet. Alkaliniteten som analyserades i processvattnet var mycket låg och till det tredje testet ökades både alkalinitet och fosfat. Detta test gav de bästa resultaten av nitrifikationstesterna och resulterade i att ammonium omvandlandes till nitrit, dock så misslyckades testet med att omvandla nitrit till nitrat. Resultatet från denitrifikationstestet visade en denitrifierande aktivitet när nitrat omvandlades till nitrit samtidigt som totalkväve minskade, vilket indikerar att kvävgas har bildats, dock borde testet ha utförts under en längre tidsperiod. Ytterligare labbanalys av processvatten från tunnlar med hög kvävehalt är viktigt vilket borde fokusera på metallkoncentrationen och proportionerna av slam och processvatten eftersom dessa parametrar inte grundligt analyserats i denna studie. Nyckelord: kväve, nitrifikation, denitrifikation, processvatten, tunnlar, inhibiterande faktorer, vattenanalys III
Chalmers University of Technology
PREFACE This master thesis has been conducted in the master’s program Infrastructure and Environmental Engineering for Skanska Sweden AB - Infrastructure with the aim of assessing nitrogen treatment during tunnel construction. The thesis was carried out from January to May 2014 with guidance from both Skanska Sweden AB and the department Water Environment Technology at Chalmers University of Technology. I would like to thank my supervisor Britt-Marie Wilén at Chalmers University of Technology for taking the time to help and guide me through the laboratory work. I would like to acknowledge my supervisors at Skanska Sweden AB, Mikael Andersson and Jenny Karlsson. Thank you for giving me the opportunity to conduct my thesis at Skanska Sweden AB and for helping me to select this topic for my thesis. I would also like to acknowledge John Nyberg at Skanska Sweden AB for your optimism, guidance and interest in this topic, thank you. This thesis would not have been completed without the help of Magnus Hallberg at Halfor AB, thank you for your ebullient commitment, guidance and competence throughout this time. Further, I would like to thank the “water boy” Niclas Johansson at Skanska Sweden AB in the project E6 Pålen - Tanumshede. Thank you for all the help with water collection, water samples and analysis at site. I would also like to acknowledge Mats Bäckman and Charlotte Bougart at Hammargårdens treatment plant. A final thank to my opponents Erik Eidmar and Johan Hultman for your valuable input and feedback on my work. This project has been very interesting and it has contributed to increase my knowledge about the complexity of water contaminants and treatment processes which is a worldwide issue. It is important to keep our waters clean, today and tomorrow, for us and our next generations. Göteborg, May 2014 Victoria Liljedahl CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: VI
Chalmers University of Technology
1 INTRODUCTION Water is an essential part of life and only 2.5 percent of water resources around the world are available freshwater. Sweden is one of few countries in the world where freshwater availability currently is not a big problem. Sweden is covered by approximately 10 percent of inland waters in terms of rivers, streams and lakes (SCB, 2012). There are 119 main watercourses with catchment areas bigger than 200 km2 that end up by the coast. There are in total 27,663 watercourses and the total length of all watercourses is 192,000 km (SMHI, 2010). Hence, surface water is not a rarity in the Swedish landscape. But even with sufficient groundwater and surface water, a sustainable management of the freshwater resources is still very important. Especially for future generations as the aim for Swedish waterways is to be ecologically sustainable (Naturvårdsverket, 2014). Due to this high areal coverage of surface water, the development of infrastructure can be challenging and complex. Infrastructural construction in or near water in Sweden is therefore inevitable. Tunnels are one construction that enables easy access to and from different locations for commute users in the Swedish landscape, both in terms of crossing waterways but also through rock passes. There is, for example, the Tingstad tunnel in Gothenburg that crosses Göta Älv, the Northern Link in Stockholm and the controversial railway tunnel through Hallandsås all of which are developed to increase the infrastructure efficiency. Tunnel construction is increasing in Sweden and there are currently around 20 road tunnels (Vägverket, 2001), which are mostly concentrated to Stockholm, Gothenburg and the West coast and there are also 168 railway tunnels around the nation (Trafikverket, 2014). There are many types of tunnel constructions in Sweden but the most common and least expensive is a rock tunnel executed by blasting (Vägverket, 2001) where most explosives used today contain ammonium nitrate, sodium nitrate and calcium nitrate as the oxidizing agent (Forsyth, et al., 1995). One issue with the explosives is the significant required concentrations of nitrogen (N) which is transferred or leached through the runoff to recipients near the tunnel construction (Forsyth, et al., 1995). The anthropological impact on the tunnel construction area, which leads to excessive nitrogen concentrations in the runoff water, could if untreated have fatal impacts on the aquatic life as well as declined quality of freshwater sources. With the large areal coverage of waterways in Sweden, the risk of contaminating the water sources during construction work is high and need cautious handling, such as efficient water treatment at the construction site. The water treatment technologies in operation today do not always meet the requirements of different parameters, especially treatment of nitrogen compounds during colder temperatures. 1.1 BACKGROUND Skanska Sweden AB is currently running a highway project in western parts of Sweden called E6 Pålen-Tanusmhede. This project started in 2013 and aims to expand and relocate the last phase of the highway E6 which runs from Pålen to Tanumshede. The new stretch of E6 in that area requires a tunnel consisting of two tubes which is performed by blasting. Blasting processes requires explosives which contain nitrogen, and the contaminants in particular are ammonium and nitrate. Nitrogen has become an issue due to its high concentrations in the water runoff from the tunnel construction which mainly leads to water contamination in two ways. The first contamination source is in the front of the tunnel where water is used to wash the blasted rock CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 1
Chalmers University of Technology
masses in order to settle dust and remove blasting residues. The other contamination source is at the disposal site where crystalline nitrogen residues are kept which is infiltrated to the groundwater or transported to the nearby recipients during precipitation. The area has suffered from quite heavy precipitation which has increased the concentration of nitrogen in the runoffs, thus putting higher pressure on the water treatment facility. The biological treatment for nitrogen removal during the project has not worked, which was presumed to be related to low water temperatures. However, other inhibiting parameters could cause the nitrogen removal failure. 1.2 AIM The aim of this master thesis is to assess which factors and parameters that affect the biological treatment of process water in tunnel construction with regard to nitrification and denitrification. The tunnel project Gerumstunneln in Tanumshede, Sweden, is used as a case study due to arisen difficulties of meeting acceptable nitrogen concentration in the recipients. The wastewater is a combination of process water from the tunneling construction, stormwater runoff and leachate water, which becomes contaminated by excessive nitrogen concentrations from explosive residues during blasting and storage. One issue is the water temperature which is below 10°C at which nitrification and denitrification processes are significantly decelerated, which limits the nitrogen removal rate. Thus, the aim is to achieve a complete nitrification and denitrification process in the lab by using the process water from the project. 1.3 SCOPE This master thesis is carried out for Skanska Sweden AB due to heightened restrictions of nitrogen concentration in process water discharge to recipients during tunnel constructions. The thesis first presents background information regarding nitrogen removal, conventional wastewater treatment and the case study. This is followed by lab analysis descriptions and its results, which in turn is used for calculations of which scale the water treatment facility need to have for different temperatures. The thesis finalizes with a discussion and a conclusion which provides necessary information to achieve sufficient nitrogen removal. 1.4 METHOD This master thesis comprises of two parts. The first part consists of a literature study with information about the current wastewater treatment method in Tanumshede and other technologies that could potentially be of relevance for tunnel constructions. The theory behind nitrogen is also presented in order to fully understand the different toxic compounds. The second part is a lab analysis where water samples from Tanumshede will be used to characterize the process water’s composition and conduct some smaller experiments to assess if the process water can be nitrified and denitrified with activated sludge from a nutrient removal wastewater treatment plant. 1.5 LIMITATIONS The primary focus will be on nitrogen compounds even though other parameters, such as heavy metals or phosphorus, could be factors that could inhibit the nitrification and denitrification in process water from other tunnel constructions. The nitrogen compounds in focus are ammonium nitrogen, nitrite nitrogen and nitrate nitrogen. Throughout the report when discussing ammonium, nitrite and nitrate, they are always referred to the forms of ammonium nitrogen , nitrite nitrogen and 𝑁𝐻4 −𝑁 𝑁𝑂2 −𝑁 CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 2
Chalmers University of Technology
nitrate nitrogen , not to be mixed with , or . Thus, if ammonium is mentioned, it is in the form of ammonium nitrogen, otherwise it will be clearly stated. 𝑁𝑂3 −𝑁 𝑁𝐻4 𝑁𝑂2 𝑁𝑂3 The lab analysis will be conducted in a smaller scale and cannot fully represent the actual treatment process in full-scale, which is more complex. The lab is also performed during a limited time which could give different results compared to studies under longer periods. The sludge needs to be collected outside of Gothenburg which also limits the time and day for when the lab can be carried out. Due to this geographical limitation, the sludge will be kept for a few days between some experiments and will not always be used the same day as it was collected from the treatment plant. There are geographical limitations for the process water as well since the tunnel project is located approximately 135 km from Gothenburg. The process water will most likely not contain much biological material why the same water will be used throughout all experiments, but it will be stored in a cool place in between. The focus in the lab will only be on nitrification and denitrification with some alterations between the different tests in order to obtain different results. There are many factors that could contribute to inhibition of nitrification and denitrification, however, since the time period for this thesis is limited, only a few factors will be analyzed. The water temperature could affect the nitrogen treatment and in reality, the temperature is around 10°C in the tunnel water. However, the tests in the lab are carried out in room temperature to observe if nitrification and denitrification could be achieved at all and to assess other possible factors that could decline the nitrifying and denitrifying activity. CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 3
Chalmers University of Technology
2 NITROGEN THEORY The largest source of nitrogen on the planet is the atmosphere which contains around 78 percent of nitrogen. Plants are dependent on nitrogen for growth and development and the main source in soils are organic matter, which in turn originates from animal and plant residues. However, plants need inorganic forms of nitrogen, thus bacteria in the soils convert organic nitrogen to inorganic forms before being taken up by the plant roots. Humans and animals eat the plants which are later returned to the soil in residues (Killpack & Buchholz, 1993). There is a deficit of nitrogen in soil which slows down the decomposition process of dead organic substances, which in turn decrease the emissions of greenhouse gases (Golubyatnikov, et al., 2013). The nitrogen cycle is presented in Figure 2.1. This shows that atmospheric nitrogen is transported to the soils and plants. Animal residues decompose to the nitrogen compound ammonium which is nitrified to nitrites and thereafter nitrates before denitrified back to the atmosphere (U.S. EPA, 2012). The nitrification and denitrification process are further described in Section 2.2.1 and 2.2.2. Rain contributes atmospheric nitrogen when it reaches the soil and could reach the groundwater aquifers. Nitrogen from factories producing commercial nitrogen fertilizers could also enter the nitrogen cycle through farming and gardening. Figure 2.1. The nitrogen cycle (U.S. EPA, 2012) In tunnel construction it is important to monitor the discharge water from the construction process. There are many parameters that should be carefully observed such as total nitrogen, pH, alkalinity, sulphite, nitrate, nitrite, ammonium, suspended particles and conductivity (Trafikverket & Ramböll, 2012). However, many parameters are treated to designated limits and do often not constitute a problem. The exception is nitrogen contaminations which is the primary nutrient of focus in this master thesis due to high concentrations in the runoffs during tunnel construction. Nitrogen is essential for growth of microorganisms, plants and animals and also in the CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 4
Chalmers University of Technology
synthesis of protein. The principle sources of nitrogen compounds in wastewater are plant and animal origins, sodium nitrate and atmospheric nitrogen. Most nitrogen compounds found in soil and groundwater are biologically originated (Tchobanoglous, et al., 2004) but during tunnel construction, excessive nitrogen compounds originate from the explosives. 2.1 NITROGEN COMPOUNDS There are several nitrogen compounds of importance but the most common in general are ammonia (NH ), nitrite (NO -) and nitrate (NO -) (Shrimali & Singh, 2001). The 3 2 3 most common compounds in wastewater are ammonia, ammonium, nitrogen gas, nitrite ions and nitrate ions (Tchobanoglous, et al., 2004). Table 2.1 summarizes some of the most common nitrogen compounds where the primary compounds in tunnel construction are ammonium, nitrite and nitrate. Table 2.1. Different forms of nitrogen and their definitions, modified from (Tchobanoglous, et al., 2004) Forms of nitrogen Abbreviation Definition Ammonia gas Ammonium ion 𝑁𝐻3 𝑁𝐻3 + + Total ammonia nitrogen 𝑁TA𝐻N4 𝑁𝐻4 + Nitrite 𝑁𝐻3 +𝑁𝐻4 − − Nitrate 𝑁𝑂2 𝑁𝑂2 − − Total inorganic nitrogen T𝑁I𝑂N3 𝑁𝑂3 + − − Organic nitrogen Organic N 𝑁TK𝐻N3 +– (𝑁𝐻4 +𝑁𝑂2 )+ 𝑁𝑂3 + Total Kjeldahl nitrogen TKN Organic 𝑁N𝐻3+𝑁𝐻4 + Total nitrogen TN Organic N+𝑁𝐻3+𝑁𝐻4 + − − Nitrogen removal is correlated to temperature where +th𝑁e𝐻 r3 e+ac𝑁ti𝐻o4 n +ra𝑁te𝑂s2 in+c𝑁re𝑂a3 se with temperature. Chang, et al., (2013) showed how the nitrogen removal efficiency changed during the four seasons for nitrate ( ) and total nitrogen (TN) which turned out to be as expected, the highest removal −efficiency during summer and the lowest during winter using the IVCW (inte𝑁gr𝑂a3 ted vertical constructed wetland) system. Vertical constructed wetlands can have up-flow or down-flow and IVCW combines these two wetlands to improve the water quality. Wetlands consist of plants and some sort of media, e.g. gravel, where the wastewater runs through the media (wetland) (Peng, et al., 2014). It is important to separate ammonium with ammonium nitrogen for example in terms of concentrations. It is the same nitrogen compound, but when concentrations are analyzed, it differs between the two types. If a concentration is presented as ammonium, a factor is required to convert this concentration to ammonium nitrogen. The atomic mass for nitrogen is 14, hydrogen 1 and oxygen 16. This is used to calculate the factors for each nitrogen compound. These factors are presented in Equations 2.1-2.4. CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 5
Chalmers University of Technology
(Eq 2.1) 14 14 𝑁𝐻3 → 𝑁𝐻3 −𝑁:1×14+(3×1) = 17 = 0.82 (Eq 2.2) 14 14 𝑁𝐻4 → 𝑁𝐻4 −𝑁:1×14+(4×1) = 18 = 0.78 (Eq 2.3) 14 14 𝑁𝑂3 → 𝑁𝑂3 −𝑁:1×14+(3×16) = 62 = 0.23 (Eq 2.4) 14 14 𝑁𝑂2 → 𝑁2𝑂.12.−1 𝑁:N1I×TR14I+T(2E× 1 6) = 46 = 0.3 Nitrite is considered to be quite unstable and is easily oxidized to nitrate. Nitrite is measure−d colorimetrically and it is an indicator of past pollution. Nitrite is often found i𝑁n 𝑂sm2 all concentrations, rarely above 1 mg/l in wastewater, but it is none the less very toxic to fish and other aquatic animals. If nitrites are found in wastewater, they are oxidized with chlorine in wastewater treatment plants in order to be reduced (Tchobanoglous, et al., 2004). If nitrite is present in high concentrations in water and if the water is ingested by human or animals, then nitrite can turn haemoglobin to methemoglobin which can cause anoxia and death since methemoglobin cannot transport oxygen to cells (Alonso & Camargo, 2003), also known as the blue baby syndrome. 2.1.2 NITRATE The most oxidized form of nitrogen in wastewater is nitrate of which the concentrations could be measured using specific-ion electrodes, o−r as nitrite with colorimetric methods. The U.S. EPA limits nitrate in drinking w𝑁a𝑂te3 r to 45 mg/l and the wastewater effluent concentration often varies between 0 and 20 mg/l. High concentrations of nitrates in water can have serious and even fatal effects for infants (Tchobanoglous, et al., 2004). Camargo et al. (2005) shows that nitrate can be removed by the use of for instance aquatic plants, algae and bacteria but also by using ion exchange, reverse osmosis (RO) and electro-dialysis since the conventional processes such as chlorination, UV and filtration are not viable for nitrate ions (Shrimali & Singh, 2001). However, the treatment processes are quite expensive and some of them rather cause new problems in terms of nitrate disposal. 2.1.3 AMMONIA AND AMMONIUM Ammonia gas is the oxidation state in most organic compounds. Ammonia can be present as either ammonia gas or ammonium ion in aqueous solutions, which depends𝑁 𝐻o3 n the pH. The concentrations of ammo+nia can be measured colometrically, titrimetrically or w𝑁it𝐻h 3 specific-ion electrod𝑁es𝐻 4 (Tchobanoglous, et al., 2004). The un-ionized molecular ammonia exists in equilibrium with the ionized ammonium dissolved in water which depends on both pH and temperature. The existence of ammonia in environments with a pH below 7 is very low, hence the ammonia ions increases with increased pH (U.S. EPA, 1993). If the concentration of ammonia in waterways is in the range of 0.2-0.5 mg/l, the toxicity to fishes can be fatal (Miladinovic & Weatherley, 2008). 2.1.4 ADVERSE EFFECTS FROM NITROGEN COMPOUNDS Three major environmental problems have been identified by Camargo & Alonso (2006) of which all include inorganic nitrogen. Eutrophication is one problem, CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 6
Chalmers University of Technology
acidification of ecological ecosystem the second problem and the third is direct toxicity which affects aquatic life’s survival, growth and reproduction. EUTROPHICATION Increased concentrations of various nitrogen forms in freshwater could generate in eutrophic conditions. Eutrophication presents itself with excessive plant growth and algae blooming which is the result of over-fertilization in water bodies. With increased plant growth, reduced DO and non-clear waters, the organisms in the water body receives less sunlight and less oxygen since it is being depleted by the growing plants and algae, hence the survival of the aquatic life could be critical (U.S. EPA, 1993). If wastewater from tunneling shows increased nitrogen values, the recipient could potentially indicate increased nutrient growth, hence degrading the water quality. Eutrophication in water resources is a worldwide problem in both freshwater and marine environment. In Sweden, the problem is concentrated in the southern parts but some mountain areas indicate eutrophication as well (Naturvårdsverket, 2013). There are many adverse effects on freshwater, some includes increased productivity and biomass of phytoplankton and suspended algae, shifts in phytoplankton composition to toxic species, threats to aquatic species, problems in drinking water supply, depletion of oxygen and decreased recreational values (Smith, 2003). It is not only the aquatic life and environmental surroundings that are affected, humans could also be threatened. When blooming algae dies, algae’s toxin can be produced which is harmful to humans. Inorganic nitrogen (ammonia, ammonium, nitrite and nitrate), in particular nitrate, is used in the eutrophication reaction but total nitrogen (TN) is also considered when assessing eutrophication. Nitrite concentration in eutrophic water is also a threat to humans since the product of nitrite nitrification is carcinogenic (Yang, et al., 2008). Nitrate could especially reduce eutrophication if it is decreased (Deng, et al., 2012). ACIDIFICATION Nitrogen dioxide and nitrogen oxide are the main nitrogen acidifying pollutants (sulphur dioxide is also a major pollutant) in lakes and streams. These pollutants emit int𝑁o 𝑂th2 e atmosphere and can𝑁 t𝑂ransform by complex reactions into sulfuric acid and n𝑆i𝑂tr2 ic acid (Baker, et al., 1991). Anthropogenic acidification of streams and lakes could cause many adverse effects e.g. biotic impoverishmen𝐻t2 o𝑆f𝑂 4 salmonids and inve𝐻rt𝑁eb𝑂ra3 tes. Other adverse affects in freshwater are increased accumulation and toxicity of aluminum (Sprenger & McIntosh, 1989), hatching delay of fish and amphibian eggs, increased migration of aquatic insect from their nests, declined zooplankton diversity and reduced growth rates in fish (Camargo & Alonso, 2006). TOXICITY Ammonia concentrations of 0.1-1.0 mg/l are often seen as lethal for the wildlife. Biological membranes are permeable to unionized ammonia which makes the toxicity of ammonia a function of pH (Abeliovich, 1992). Ammonia is very toxic to the aquatic wildlife, compared to the non- or less toxic ammonium (Camargo & Alonso, 2006). Nitrite concentrations on the other hand should not exceed 0.1 mg/l in closed water systems. However, lethal concentrations are much higher, around 250-350 mg/l proved to be lethal to many fish, hence nitrite is less toxic than ammonia but should still be kept under observations (Abeliovich, 1992). Though, other studies show that CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 7
Chalmers University of Technology
nitrite concentrations above 45 mg/l could result in anemia in infants and pregnant women (Odjadjare & Okoh, 2010). The main toxic effect of nitrite on fish is hypoxia which later results in death. This is due to a conversion of pigments former carrying oxygen shifting to a state where they are incapable o carrying oxygen. Other effects of nitrite on fish are cause of electrolyte imbalance, effects on membrane potentials, skeletal muscle contractions and heart function, tissue shortage and repression of the immune system (Jensen, 2003). 𝑂2 Nitrate toxicity has been considered irrelevant since it must be converted to nitrite before it becomes toxic. The nitrate uptake is much more limited than nitrite uptake which contributes to low nitrate toxicity. However, studies have shown that small concentrations of nitrate can have adverse effects on sensitive aquatic animals during long-term exposure. Further, freshwater animals tend to be more sensitive to nitrate toxicity compared to marine animals (Camargo, et al., 2005). When assessing the effects of increased nitrogen concentration in drinking water of humans, methemoglobin could be the result from conversion of nitrates to nitrites under anaerobic conditions in the digestive tract (Greer & Shannon, 2005). Symptoms are headache, fatigue, stupor, convulsions and could lead to death. Ingested nitrates and nitrites could be contributing factor of developing cancer of the digestive tract (Fewtrell, 2004). Adverse affects from blooming toxic cyanobacteria on humans have been reported around the world, including Sweden, with effects such as eye irritation, fever, diarrhea, skin rash and muscular cramps. Some cases of human exposure led to lethal outcomes where children are the most sensitive group ( (Hitzfeld, et al., 2000) and (Chorus, 2001)). As mentioned earlier, algae can cause toxicity to aquatic animals and these toxins can either remain in the algal cells or be released into the water. Thus, animals can be directly exposed to toxins through absorption, drinking or ingesting by feeding activities. Cyanobacteria, diatoms and dinoflagellates are the predominate groups that contribute to toxic algae (Camargo & Alonso, 2006). 2.2 BIOLOGICAL TREATMENT FOR NITROGEN Water in tunnel constructions differ in composition compared to regular household wastewater (e.g. the lack of faecal contaminations). A conventional wastewater treatment plant consists of mechanical, chemical and biological treatment and sometimes disinfection as a final step. Conventional wastewater treatment is necessary as a pre-treatment step for nitrogen removal. The biological treatment step use microorganisms such as bacteria in order to feed on the organic left-over matter from the mechanical treatment step (The World Bank, n.d). Microorganisms feed on nitrogen and around 20 percent is removed by flocculation and separation (Naturvårdsverket, 2009). However, since all nitrogen is not fully removed in biological treatment, a specific nitrogen removal step is added which is used in bigger wastewater treatment plants serving more than 10,000 people and sensitive recipients (Naturvårdsverket, 2009). Ammonia is converted into nitrate by nitrifying bacteria in an aerobic condition after which denitrifying bacteria convert nitrate into nitrogen gas in anoxic conditions (Gryaab, n.d). These nitrifying and denitrifying bacteria can be obtained from sludge in a wastewater treatment plant. The anticipated nitrogen removal efficiency is around 50 to 75 percent (Naturvårdsverket, 2009). Nitrification and denitrification processes are the basis of all biological nitrogen removal methods at the moment used for wastewater treatment. This has proved to be CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 8
Chalmers University of Technology
both economically and technically feasible for centralized and decentralized systems. Aerobic processes (oxygen rich) nitrify ammonium, i.e. transform ammonium into nitrate where after anoxic processes (oxygen absent) reduce nitrate to nitrogen gas (Oakley, et al., 2010). Ammonium and organic nitrogen are the most common compounds of nitrogen in a modern wastewater treatment plants and is removed by biological nitrification/denitrification (Van Hulle, et al., 2010). Figure 2.2 presents three different nitrification and denitrification processes where a) is a pre-anoxic process starting with denitrification in an anoxic tank with wastewater as the carbon source and is followed by nitrification in the aeration tank. The nitrified effluent is Figure 2.2. Nitrification and denitrification processes, retrieved from Oakley later recycled back to the anoxic tank. In process b), which is a post-anoxic process, the order is reversed from the pre-anoxic system and the carbon source can be either external sources such as methanol or endogenous respiration of bacterial cells. In c) the nitrification and denitrification occurs simultaneously in the same reactor (Oakley, et al., 2010). 2.2.1 NITRIFICATION The nitrification process is presented in Reaction 2.1 and 2.2 which describe the two- step biological process of ammonia oxidized to nitrite followed by oxidation of nitrite into nitrate (Tchobanoglous, et al., 2004). Reaction 2.1 shows how ammonia (or ammonium) is converted into nitrite with ammonium-oxidizing bacteria (AOB) such as Nitrosomonas as the most common group of bacteria while Reaction 2.2 converts nitrite into nitrate by nitrite-oxidizing bacteria (NOB) such as Nitrobacter (U.S. EPA, 2002). (Re 2.1) − + − (Re 2.2) 𝑁𝐻3 +𝑂2 → 𝑁𝑂2 +3𝐻 +2𝑒 − − + − Reactions using the AOB-bacteria and NOB-bacteria can also be presented as 𝑁𝑂2 +𝐻2𝑂 → 𝑁𝑂3 +2𝐻 +2𝑒 Reaction 2.3 and 2.4 in the nitrification process which nitrifies ammonium to nitrite and nitrate (Tchobanoglous, et al., 2004). CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 9
Chalmers University of Technology
(Re 2.3) + − + (Re 2.4) 2𝑁𝐻4 +3𝑂2 → 2𝑁𝑂2 +4𝐻 +2𝐻2𝑂 − − The total oxidation reaction is presented in Reaction 2.5 which shows how ammonium 2𝑁𝑂2 +𝑂2 → 2𝑁𝑂3 is oxidized into nitrate (Tchobanoglous, et al., 2004). (Re 2.5) + − + The nitrification process is not able to remove the toxic nitrogen compound ammonia, 𝑁𝐻4 +2𝑂2 → 𝑁𝑂3 +2𝐻 +𝐻2𝑂 but the process converts ammonia into nitrate as stated in the reactions presented earlier. This reduces or eliminates the toxicity to fish as well as reducing the nitrogen oxygen demand (NOD) of the effluent. The rate of the nitrification process in wastewater is a function of time, and also, it is independent of the ammonia nitrogen concentration (Viessman, et al., 2009). A low growth rate characterizes nitrifying bacteria which is due to the low energy yield, which in turn is connected to the oxidation of ammonia and nitrite. Most nitrifying bacteria are autotrophic and use carbon dioxide as the carbon source, which should be reduced before the carbon reacts with the cell mass. The reduction is performed through oxidation of ammonium and nitrite hence the reaction for growth is shown in Reaction 2.6 and 2.7 (Henze, et al., 1997). (Re 2.6) + − + (Re 2.7) 15𝐶𝑂2+13𝑁𝐻4 → 10𝑁𝑂2 +3𝐶5𝐻7𝑁𝑂2 +23𝐻 +4𝐻2𝑂 + − − + Equation 2.5 is a suggested expression by the U.S. EPA (1993) of the maximum 5𝐶𝑂2 +𝑁𝐻4 +10𝑁𝑂2 +2𝐻2𝑂 → 10𝑁𝑂3 +𝐶5𝐻7𝑁𝑂2 +𝐻 growth rate of Nitrosomonas as a function of temperature over the range 5-30°C. (Eq 2.5) 0.098(𝑇−15) 𝜇𝑁 = 0 .47𝑒 −1 𝜇𝑁 = 𝑚𝑎𝑥𝑖𝑚𝑢𝑚 𝑠𝑝𝑒𝑐𝑖𝑓𝑖𝑐 𝑔𝑟𝑜𝑤𝑡ℎ 𝑟𝑎𝑡𝑒 𝑜𝑓 𝑁𝑖𝑡𝑟𝑜𝑠𝑜𝑚𝑜𝑛𝑎𝑠 𝑑 𝑒 = 𝑏𝑎𝑠𝑒 𝑜𝑓 𝑁𝑎𝑝𝑖𝑒𝑟𝑖𝑎𝑛 𝑙𝑜𝑔𝑎𝑟𝑖𝑡ℎ𝑚𝑠,2.718 Since the temperature is affecting the nitrification, winter aeration need to be much 𝑇 = 𝑡𝑒𝑚𝑝𝑒𝑟𝑎𝑡𝑢𝑟𝑒,°𝐶 longer than during summer. There are some ways to solve those seasonal issues, the mixed-liquor suspended solids (MLSS) could be increased and pH could be adjusted (Viessman, et al., 2009). 2.2.2 DENITRIFICATION The denitrification process reduces the oxidized forms of nitrogen to nitrogen gas by heterotrophic bacteria and by using a carbon source of biodegradable organic matter (Mees, et al., 2014). First, there is a biological reduction of nitrate to nitric oxide NO, then nitrous oxide and finally nitrogen gas N (Tchobanoglous, et al., 2004). The end 2 product, nitrogen gas, has no significant effect on the environment (U.S. EPA, 1993). Denitrification occurs when oxygen is depleted, hence the process only occurs in anaerobic or anoxic environments with enough quantities of nitrate and where the oxygen demand surpasses the oxygen supply (Deng, et al., 2012). The nitrification process has limited options for types of bacteria whilst denitrification has a broad range of useful bacteria, which can be both heterotrophic and autotrophic. Some heterotrophic organism genera are Achromobacter, Bacillus, Flavobacterium and Psedumonas, where the latter is the most common denitrifier. Most of the heterotrophic bacteria have the ability to use oxygen, nitrite or nitrate CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 10
Chalmers University of Technology
(Tchobanoglous, et al., 2004) though nitrite or nitrate is the wanted forms for the denitrification process. However, if oxygen is present it will be used naturally instead of nitrate and vice versa in anoxic environments. Nitrite and nitrate are considered to be the electron acceptors in the respiratory electron chain, which is the primary mechanism for energy generation in cells. Consequently, nitrite and nitrate replaces oxygen in the transport chain which generates to some extent less energy (U.S. EPA, 1993). A simplified denitrification reaction is seen in Reaction 2.8 (Henze, et al., 2002). (Re 2.8) − − The denitrification process needs an organic carbon source to act as a hydrogen donor 𝑁𝑂3 → 𝑁𝑂2 → 𝑁𝑂 → 𝑁2 and to supply the biological synthesis. Some organic carbon sources could be acetic acid, ethanol, methanol or organic matter. Methanol is preferred common source since it is the least expensive synthetic compound that can be used due to its quality of not leaving residual BOD in the effluent. Methanol first reduces DO in order for the biological reduction of nitrate and nitrite to take place (Viessman, et al., 2009). These different reactions, using methanol ( ) are presented in Reactions 2.9-2.11. (Re 2.9) 𝐶𝐻3𝑂𝐻 (Re 2.10) 3𝑂2 +2𝐶𝐻3𝑂𝐻 = 2𝐶𝑂2 +4𝐻2𝑂 − − (Re 2.11) 6𝑁𝑂3 +5𝐶𝐻3𝑂𝐻 = 3𝑁2 +5𝐶𝑂2 +7𝐻2𝑂+6𝑂𝐻 − − The energy yielding process for denitrifying bacteria in Reaction 2.12 is a 2𝑁𝑂2 +𝐶𝐻3𝑂𝐻 = 𝑁2 +𝐶𝑂2 +𝐻2𝑂+2𝑂𝐻 combination of two half-expressions. The process uses organic matter in wastewater as the energy and carbon source (Henze, et al., 1997). (Re 2.12) 1 1 3 1 + 70𝐶18𝐻19𝑂9𝑁+5𝑁𝑂3 +5𝐻 → 1 17 1 − 1 + 1 𝑁2 +2.2𝐶.𝑂32 +TRE𝐻A𝐶T𝑂M3 E+NT 𝑁T𝐻IM4 +E SP𝐻A2𝑂NS 10 70 70 70 5 The trial spans are quite long for many of the researches in the literature with examples of observation and analyze time of 180, 225 and 550 days for nitrification and denitrification processes (Hoilijoki et al. 2000, Jokela et al. 2002 and Koren et al. 2000). Isaka et al. (2007) had a start-up period of 3 months before the nitrification process was fully working whilst Jokela et al. 2002 had less than 3 weeks of start up for nitrification and around 2 weeks for denitrification, though the entire analysis of nitrification and denitrification proceeded over more than 225 days. Thus the time for nitrification and denitrification can vary significantly. 2.2.4 NITRIFICATION AND DENITRIFICATION RATES There are different nitrification and denitrification rates presented in several studies, summarized in this section and in Table 2.2 and 2.3. One research using cold mine water had nitrification rates of 29 and 38.3 g N/m3 at 5°C and 10°C respectively. The nitrification rate at 5°C increased to 43 g N/m3 if the salinity was increased. The denitrification rates were 605 and 522 g N/m3 for 5°C and 10°C (Karkman, et al., 2011). Zaitsev et al. (2008) had a nitrification removal rate of 0.54 g NH -N/m2-d as 4 the highest at 5°C. The denitrification rate was 4.1 g NO₃-N/m2,d at 5°C. There were also presented other denitrification rates at 3°C and 7°C of 1.5 g NO₃-N/m2-d and 1.6 g NO₃-N/m2-d respectively. CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 11
Chalmers University of Technology
One study performed by Rostron et al. (2001) achieved different nitrification rates for CSTRs (continuously stirred tank reactors) which varied from 2.28 to 4.24 g N/m2- d,media with the synthetic ammonia feed with a concentration of ammonia nitrogen 500 mg/l. There are some other nitrification rates found in the literature using immobilized biomass. With the use of domestic sewage for example, a rate of 0.29 kg N/m3-d was obtained with the ammonia nitrogen concentration of 50 mg/l. Several studies have used synthetic feed water where the rates have been 0.58, 0.2, 0.25, 0.73 and 1.5 kg N/m3-d with respectively 27, 35, 50, 200 and 50 mg/l of ammonia nitrogen (Rostron, et al., 2001). Hoilijoki et al. (2000) reported maximum nitrification rates of 0.13, 0.27±0.06 and 1.78 mg NO -N g/VSS,h.at 7, 10 and 24°C in December, August 3 and July respectively. This study was conducted using an aerobically pretreated leachate. Some in-situ nitrification rates during cold seasons (average of 10°C) varied for different sites from 204 to 548 mg N/m2,d while the denitrification rates in the same season varied from 180 to 491 mg N/m2,d. These rates are based on influent values of 62 mg/l for TN and 60 mg/l for NH -N. Sewage with low median strength was used in 4 the study for colder temperatures (Zimmo, et al., 2004). The nitrification rate in sediment in lakes have been assessed which showed to be much higher than in the water columns. The ammonia nitrogen concentration in the sample was 2-10 mg/l and the nitrification rates in the sediments of two different lakes resulted in 0.32 and 0.37 g N/m2-d (Pauer & Auer, 2000). The nitrification rate was analyzed in leachate from a landfill in Japan which had the characteristics of 400-610 mg/l NH -N and 420-650 mg/l TN in the influent water. 4 The effluent water had NH -N concentrations of 16-35 mg/l which is a large decrease. 4 The highest nitrification rate was 0.71 kg N/m3-d at 10°C which took up to 3 months to achieve (Isaka, et al., 2007) while Jokela et al. (2002) achieved nitrification rates of 0.05 kg N/m3-d at 5 and 10°C. Table 2.2. Nitrification rates.1) (Kar kman, et al., 2011), 2) (Rostron, et al., 2001), 3) (Isaka, et al., 2007), 4) (Jokela, et al., 2002), 5) (Zimmo, et al., 2004) and 6) (Pauer & Auer, 2000). Nitrification Temp Rate Unit Type of Temp Rate Unit Type of °C °C water water 5 291 g N/m³ Mine water 5/10 504 g N/m³-d Leachate water 5 381 g N/m³ Mine water - 2902 g N/m³-d Syn. wastewater 10 431 g N/m³ Mine water 10 0,2045 g N/m²-d Wastewater - 5802 g N/m³,d Syn. wastewater 10 0,5485 g N/m²-d Wastewater - 2002 g N/m³,d Syn. wastewater - 0,326 g N/m²-d Lake sediment - 2502 g N/m³,d Syn. wastewater - 0,376 g N/m²-d Lake sediment - 7302 g N/m³,d Syn. wastewater - 2,282 g N/m²-d,med Syn. wastewater - 15002 g N/m³,d Syn. wastewater - 4,242 g N/m²-d,med Syn. wastewater 10 7103 g N/m³,d Leachate water CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 12
Chalmers University of Technology
Table 2.3. Denitrification rates. 1) (Karkman, et al., 2011), 2) (Zimmo, et al., 2004) and 3) (Zaitsev, et al., 2008) Denitrification Temp °C Rate Unit Type of water 5 5221 g N/m³ Mine water 10 6051 g N/m³ Mine water 10 0,182 g N/m²,d Wastewater 10 0,4912 g N/m²,d Wastewater 3 1,53 g NO₃-N/m²,d Mine water 5 4,13 g NO₃-N/m²,d Mine water 7 1,63 g NO₃-N/m²,d Mine water 2.2.5 SLUDGE FOR BIOLOGICAL TREATMENT The sludge concentration varies throughout the literatures in different nitrification and denitrification studies. The study performed by Koren et al. (2000) used 750 ml liquid from activated sludge reactors which was mixed with 16 liters of water, thus around 4.5 percent of the total sample. The aeration flow was 2.4 l/min and the agitation rate was 1 000 rpm. Rostron et al (2001) used nitrifying biomass of 2.5 g VSS/l using CTSR (continuously stirred tank reactors) of 4 liters. Both Jokela et al. (2002) and Hoilijoki et al. (2000) used nitrifying activated sludge of 4 g VSS/l when studying leachate. Another study on wastewater in Hong Kong used a sludge concentration of 2 g MLVSS/l (mixed-liquid volatile suspended solids) in combination with aeration of 8 l/min (Li, et al., 2013). 2.2.6 INHIBITING FACTORS Nitrification and denitrification can be inhibited for many reasons. Both occurrence and rate of the nitrification process are controlled by environmental and operating conditions, e.g. temperature and sludge age (Viessman, et al., 2009) and other parameters such as phosphate concentration and pH have also been proved to limit the nitrification and denitrification processes. Some of these factors will be considered during the lab analysis. TEMPERATURE Nitrification and denitrification processes are strongly dependent on the temperature where the rates of nitrification could double with a temperature increase of 10°C while the denitrification process could double for every increase of 4°C (Zaitsev, et al., 2008). Koren et al. (2000) states that nitrification is affected in temperatures below 10°C since the bacterial metabolism decreases and at temperatures below 4°C, the bacterial metabolism stops all together. However, nitrogen removal is achievable at temperatures below 10°C. The reduction rate is though decreased compared to room temperature of 20°C. One study showed a 30 percent nitrification reduction at 10°C compared to that of 20°C (Alawi, et al., 2009). Jokela et al. (2002), Hoilijoki et al. (2000) and Zaitsev et al. (2007) showed that a nitrification efficiency of higher than 95 percent could be achieved with different loadings. However, other literature found a 100 percent reduction in nitrification at 5°C and pointed out that high nitrification CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 13
Chalmers University of Technology
efficiency in low temperatures can only be achieved with long retention times or a combination of increased nitrifier concentration and elevated temperature if it is possible to heat the water (Ducey, et al., 2010). The bacteria genera Nitrospira is favoured during cooler temperatures and increased dissolved oxygen (Wells, et al., 2009). Adapting the nitrification process to lower temperatures has been an issue in colder northern regions such as Sweden. Some techniques have been tried and proved to be successful at 14°C and even at 7°C. They include extension of the aeration period and supplemental additions of nitrifying biomass in an aeration tank (Ward, et al., 2011). The nitrification rate decreases with about 50 percent for every 10-12°C drop in temperatures above 10°C and when the wastewater is even colder, e.g. decreasing the temperature to 5°C from 10°C, the ammonia oxidation is halved (Viessman, et al., 2009). NITROGEN COMPOUNDS Nitrite-nitrogen can inhibit the biological process if the concentration is too high. Nitrite-nitrogen concentration of 150 mg/l proved to inhibit the denitrification process while 1 350 mg/l of nitrate-nitrogen inhibited the denitrification process. However, inhibition of denitrification is not only dependant on nitrite concentration but the biomass concentration. In an experiment using the same nitrite-nitrogen concentration of 20-25 mg/l and a pH of 7-8, the denitrification was only completed with a biomass concentration of 500-1 000 mg/l whilst using a biomass concentration of 100-150 mg/l, the denitrification was inhibited. It was also concluded in this study that a decreased pH generates increased inhibition (Glass, et al., 1997). A critical nitrite- nitrogen dose for nitrification is 50-100 mg/l at a pH 7 which caused inhibition. Moreover, the inhibition lasted more than 10 days after the nitrite disappeared from the solution (Philips & Verstraete, 2000). Ammonia, i.e. molecular free ammonia, is the major compound responsible for toxicity effects on aquatic life (U.S. EPA, 1993) and it can suppress the nitrifying process. Toxicity can have great impacts on nitrifiers and can kill the nitrifying bacteria. Free ammonia can halt the nitrification process in terms of being inhibiting on the bacteria Nitrosomonas at free ammonia concentrations of 10-150 mg/l and at 0.1-1 mg/l for Nitrobacter (U.S. EPA, 1993). Other studies show that concentrations of free ammonia higher than 0.2-1 mg/l inhibit the nitrite oxidation (He, et al., 2012). PHOSPHATE Phosphorus is most often present in the environment as organic or inorganic phosphate. Phosphate has been reported to limit nitrification, especially during low temperatures. In some nitrifying systems using filters, the phosphate concentrations were dosed 5-50 µg (de Vet, et al., 2012). de Vet et al. (2012) performed a study using filters. The3 −two different tests using filters with 80 µg had complete nitrification𝑃 𝑂pr4 oce−ss𝑃es whilst the two other tests with 20 and 30 µg resulted in incomplete nitrification. However, with concentration around𝑃 𝑂104 0− µ𝑃g, the nitrification was inhibited as well. 𝑃𝑂4−𝑃 One pilot study by Kors et al. (1998) showed that by dosing with phosphate, the ammonium removal increased in the nitrification process. With phosphate concentrations below 15 µg /l in combination with very low temperatures, the nitrification process failed to convert ammonium to nitrate. It was then assumed that the phosphate concentration w𝑃a𝑂s 4 a limiting factor for ammonium removal besides the temperature. When the influents in that study were dosed with 100-500 µg /l, 𝑃𝑂4 CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 14
Chalmers University of Technology
ammonium removal rates increased despite temperatures around 1°C (Kors, et al., 1998). Another pilot study also proved that phosphate was limiting the nitrification process. When dosing10 µg , the ammonium removal rate increased and the nitrification process could be com3−pleted (van der Aa, et al., 2002). 𝑃𝑂4 𝑃 METALS Metals could also be inhibiting for the nitrification process. Some metals that have been identified as inhibitors are zinc, copper, chromium, nickel and cobalt (U.S. EPA, 1993). Some inhibiting concentration ranges for certain metals have been identified such as 0.08-0.5 mg/l for zinc and 0.005-0.5 mg/l for cupper, thus cupper inhibits the nitrite and nitrate production greater than zinc (Juliastuti, et al., 2003). Juliastuti et al. (2003) showed that when the concentrations of heavy metals increased, the inhibition also increased on nitrification. For example, they concluded that a 97 percent inhibition was reached when zinc had a concentration of 1.2 mg/l. Tchobanoglous et al. (2004) also presented some concentrations at which the metals could be inhibiting; 0.25 mg/l nickel, 0.25 mg/l chromium and 0.10 mg/l copper have for example proved to completely inhibit the ammonia oxidation. When it comes to nickel and cadmium, a study showed that both nickel and cadmium hade higher inhibiting effects on the ammonium oxidation rate than on the nitrite oxidation rate (Hu, et al., 2002). pH The pH can inhibit the nitrification rate. Nitrification should be performed in conditions when the pH is around 6.5-8 but for optimal nitrification rates the pH should be 7.5 to 8.0 according to the U.S. EPA (1993). The percent of the maximum nitrification rates increases as pH increases from 5.0 to 8.0 before the maximum rates starts to decline as the pH continues to increase from 8.0. With pH-values below 6.8 the nitrification rate decline quite notably, the rates could be as low as 10 to 20 percent of the rate at pH 7 when pH decreases to around 5.8 and 6. 90 percent of the maximum rate is occurring at a pH of 7.5 and 8.5, and less than 50 percent of the rate occurs below a pH of 6.4 and above 9.6 (Viessman, et al., 2009). However, values of 7.0 to 7.2 are often used to maintain a reasonable nitrification rate. An optimal value of pH for denitrification ranges between 7.0 and 9.0 though local conditions can cause variations. If the pH decreases below 7.0 it will affect the end product since nitric oxides, such as , are produced when pH is declining (Henze, et al., 2002). However, Viessman et al. (2009) claims the optimum pH to be between 6.5 and 7.5 (in accordance with m𝑁o2 s𝑂t heterotrophic bacteria) and that the rate is decreased with up to 80 percent of the maximum if pH is reduced to 6.1 or raised to 7.9. SLUDGE AGE A research on nitrogen removal in wastewater using biological treatment showed that the sludge age is a factor that affects the biological treatment. The nitrification process was unstable in temperatures lower than 15°C and sludge age lower than 20 days. Though, when the sludge age was higher than 20 days, the colder temperatures had lower influence on the process which was considered to be stable. The sludge age effect on the nitrification process was analyzed during winter (<15°C) and summer (>15°C) conditions which resulted in that the sludge age had no influence on the nitrification process during summer while a sludge age above 20 days eliminated the negative effects of low temperatures on nitrification (Komorowska-Kaufman, et al., 2006). CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 15
Chalmers University of Technology
3 NITROGEN IN TUNNEL CONSTRUCTION The drill and blast technique is used in Gerumstunneln which is a conventional tunnel construction method used as the primary technique in Sweden (Vägverket, 2008). The drill and blast technique is carried out in cycles where each cycle contains several steps that are needed to blast a few meters of tunnel. The first step in the drill and blast technique is most often pre-grouting. Groundwater leakages are often expected during tunnel construction why pre-grouting will be conducted regularly. The next step is drilling of boreholes in the required parts of the rock which are loaded with explosives. The boreholes could be approximately 2-6 meters deep but vary depending on the project (Vägverket, 2009). The third step is the actual blasting followed by transportation of the blasted and crushed material. The last step is mucking and scaling under which loose rock is removed at the blasted surface in order to avoid any large blocks of rock to fall down. At times, extra reinforcement is required such as bolting, shotcrete and post-grouting (Grinder, 2003). When one cycle is performed, another cycle begins which is followed by more cycles until the requested length of the tunnel is blasted. The cycles can vary in time due to variations of the rock quality along the tunnel stretch. 3.1 EXPLOSIVES Both drilling and blasting contribute to dust which is managed by flushing water on the working face in order to settle the dust and reduce health risk for the workers. Dust is thus transported via the water in the tunnel and could be a potential threat for the recipient. The explosive itself during tunnel blasting is responsible for producing several bi-products which also will be mixed with water, hence polluting the wastewater further (Houfeng, et al., 2013). An explosive material can be defined as a compound that reacts very rapidly and generates large quantities of gases in conjunction with liberation of heat (Sudweeks, 1985). The nitrogen residue from blasting is easily transferred into the water from the working face. Sometimes, for practical reasons, not all explosives are fully detonated and nitrogen in the chemical forms of ammonium and ammonia contaminates water from remains of undetonated explosives (V+ikan & Meland, 2012) and nitrates can leach from the explosives in wet blast𝑁 h𝐻o4 les or during char𝑁gi𝐻ng3 (Forsyth, et al., 1995). There are different types of explosives, but ammonium nitrate is a common component. Nitrogen dioxide is produced with an ammonium nitrate explosive which can be seen in Reaction 3.1 whilst Reaction 3.2 shows how the gas , which is freely soluble, is tran𝑁s𝑂fo2 rmed into nitric acid. Ammonium nitrate is an emulsion explosive and the main pollutants are ammonia and nitrate (Ho𝑁uf𝑂e2 ng, et al., 2013). 𝑁𝐻3 𝑁 𝑂3 (Re 3.1) 3 1 𝑁 𝐻4𝑁𝑂3 → 4𝐻2𝑂+2𝑁𝑂2 (Re 3.2) 1 2 1 𝑁Sp𝑂il2la+ge 3 o𝐻f2 e𝑂xp→lo 3si𝐻ve𝑁s𝑂 d3u+rin 3g𝑁 charging and undetonated explosives accounts for most nitrogen that leaks to recipients near the construction. Blasted rock masses contain the largest part of nitrogen compounds, as much as 60-70 percent of the nitrogen leaked to water comes from these masses (Tilly, et al., 2006). Commercial and modern explosives usually contain a fuel and oxidizer and examples of oxidizing agents are CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 16
Chalmers University of Technology
ammonium nitrate , calcium nitrate and sodium nitrate . Explosives of common use can be divided into three groups; ANFO (ammonium nitrate and fuel oil𝑁),𝐻 w4𝑁at𝑂e3 rgels/slurries and e𝐶m𝑎𝑁ul𝑂si3 ons. All groups of ex𝑁pl𝑎o𝑁si𝑂ve3 s contain nitrogen compounds but with different water resistance, i.e. different solubility in water, hence varying degrees of introducing nitrogen to water systems (Forsyth, et al., 1995). Blasting work in tunnels generally requires more explosives per m3 rock compared to open pits. Around 1.5-2 kg/m3 explosives per solid m3 rock is used during tunnel construction in Sweden and around 50 000 ton of explosives are used yearly in Sweden for mining, tunneling and other activities (Tilly, et al., 2006). ANFO The explosive ANFO is constituted by a relationship of around 95 percent ammonium nitrate and 5 percent diesel oil. The nitrogen content is around 34 percent in ANFO (Grinder, 2003) and the two nitrogen compounds ammonia and nitrate ions are very water soluble forms of nitrogen. ANFO has no water resistance why nitrogen is soluble when or if exposed to water. ANFO is also hygroscopic, i.e. absorbs any available water and if ANFO absorbs too much water it becomes de-sensitized and will not detonate (Forsyth, et al., 1995). This in turns leads to explosives being left undetonated in boreholes, hence nitrogen could potentially leach to the construction water. SLURRY MIXTURES Slurry mixtures, also known as water gels or dense blasting agents, consist of sensitizer, an oxidizer, water and a thickener (Nichols Jr., 2005). Slurries contain the same soluble nitrogen compounds as ANFO but with the difference that the water resistance is good once the cross-linker has activated the gum. The gelled gum creates an impermeable barrier between oxidizing agents and external water. The nitrogen content is between 20 to 30 percent in the mixture according to Forsyth et al. (1995). EMULSION The third group of explosives is emulsion explosives (EMX) which as the two latter groups also is constituted by the water soluble compounds ammonia and nitrate. The water in oil emulsion is very resistant and there is a thin film of oil surrounding a salt solution, which limits the contact with external water. An emulsion mixture contains around 20 to 30 percent of nitrogen (Forsyth, et al., 1995). Emulsion explosives containing no secondary explosives represent most commercial explosives produced today. There are many advantages with emulsion explosives compared to other explosives such as high safety in use, low cost, lower environmental impacts and the possibility of production at the construction site thus reducing transports. The amount of toxic nitrogen from emulsion explosives are 5-15 times less compared to commercial explosives with ammonium nitrate with the secondary explosive TNT (Yunoshev, et al., 2013). Emulsion explosives contain less pure nitrogen compared to ANFO, around 33-75 percent of ANFO’s nitrogen content (Tilly, et al., 2006). The explosive used in the Gerum tunnel is an emulsion explosive called site sensitised emulsion (SSE) which become more popular in recent years due to developments of using it underground. The explosive was first introduced by Dyno Nobel in 1995 for underground blasting which is today known as the Titan SSE-system. This emulsion matrix is classified as UN 5.1 oxidizing agent, meaning that is has other restrictions regarding storage and transport compared to conventional explosives (Fauske, 2003). SSE is an emulsion, hence the name, with an aqueous nitrate solution and an oil phase CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 17
Chalmers University of Technology
which is mixed with a chemical sensitizer (Bakke, et al., 2001). The matrix only becomes explosive after a couple of minutes when it has been charged in the borehole together with additive of gas. This system often include a charging truck, two container tanks and pumps for the materials; emulsion matrix and sensitizing agent (Fauske, 2003). 3.2 WATER SOURCES IN TUNNELING There are several types of water sources during tunnel constructions that in different ways are affected by the tunnel construction in terms of excessive nitrogen concentrations. Some effects such as groundwater contamination can be crucial during the project, but also long-term if the water management is not handled properly. This section presents some different water sources in tunnel construction in order to clarify how they differ amongst each other as well as their relevance in a tunnel construction. BILGE WATER During the construction process, inflowing groundwater and stormwater will be collected in excavation shafts which can be contaminated by various pollutants from the construction site (Trafikverket, 2011). This water accumulation is called bilge water and need to be treated before reaching the recipient. DRAINAGE WATER Drainage water is groundwater that infiltrates in to the tunnel and needs to be removed in separate drainage pipes (Trafikverket, 2011). The groundwater drops from the sealed rock and are led to drainage pipes by a waterproof membrane. It can also be led to the drainage pipes through the layer of macadam in the bottom of the tunnel (Trafikverket, 2012). LEACHATE WATER Leachate is wastewater created by percolation of rainwater and moisture in landfills through different types of waste. Leachate at tunnelling sites contains high concentrations of ammonium nitrogen and organic and inorganic compounds are transported from the waste to recipients (Hasar, et al., 2009). The rock material from blasting has to be stored before it is crushed and used, thus storage areas at the project site are exposed to explosive residues which contains nitrogen compounds. Precipitation infiltrates the landfills and cause leachate water with high nitrogen concentrations. PROCESS WATER During tunnel construction in rock, large volumes of process water are generated. The process water is used during drilling of boreholes for blasting, probing and grouting, as well as in cooling processes of the augers (drills). Large volumes of water are also needed to moist blasted rock masses and to wash rock surfaces prior to using shotcrete (Trafikverket & Ramböll, 2012). The process water is often contaminated due to residues in the construction process, such as oil and explosives (Trafikverket, 2011). Thus process water needs proper managing before discharged to recipients in order to reduce the contamination risk to the environment. The process water from the Gerum tunnel will be used in the lab analysis in Chapter 4 and consist of groundwater from the area. CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 18
Chalmers University of Technology
3.3 WATER TREATMENT IN THE GERUM TUNNEL The first proposed water treatment process was introduced in August 2013 which consisted of pH adjustment, a flocculation chamber, a lamella separator, DynaSand Oxo (Figure 3.1) for nitrification, a carbon source and finally DynaSand filters (Figure 3.2) with mechanical and contact filtration as well as biologically activated filtration. But this treatment process failed to meet the demands from the Swedish Traffic Administration, in terms of nitrogen concentration. Many efforts were done trying to activate the biological treatment, but there were no successful outcomes, which is assumed to be related to low temperatures (<10°C). The sludge was collected Figure 3.1. Dyna Sand Figure 3.2. Lamella Filters (Nordic Water, 2014) Filters (Nordic Water, 2014) from a local wastewater treatment plant which could be, in combination with cold water temperatures, one reason for why the nitrogen removal process decelerated. Low water temperatures will be a reoccurring problem for tunnelling construction in Sweden since the tunnel water will at most times be below 10°C (Hallberg, 2014). Since the original water treatment design could not meet the demands of nitrogen concentrations, especially the biological treatment, the treatment process was altered. As from November 2013, the process water is now treated in fewer steps without the biological treatment, hence the current treatment focuses on suspended particles, pH and oil contaminants. This treatment is functioning as a pre-treatment for nitrogen removal, which is considered to be efficient. The process water from the tunnel construction is first led to a container with oil skimmer treatment and then via a 1 m3 container to a flocculation chamber. After the 1 m3container, a manometer is used which is followed by addition of a carbon source if pH needs adjustment. The carbon source is carbonic acid which needs proper mixing before continuing to the next step which is additive of flocculants before the coagulation of particles in the flocculation chamber. After flocculation, the water goes through a lamella separator to remove CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 19
Chalmers University of Technology
sludge. From the lamella separator, the water is transported to a new container before ending up in an artificial dam and the sludge is collected by sludge trucks. The process water is currently recycled back into the tunnel construction due to limited water extraction from the water supply in the area which is around 1.8 m3 per hour. Thus the artificial dam is used, which protects the recipient to some extent. However, construction water outside the tunnel, e.g. rainwater mixed with blasting materials in storage areas, is led to a natural dam before reaching the recipient. The natural dam shows values precisely above allowed nitrogen concentrations. The main idea is to let all water go back to the recipient but since the nitrogen levels have been too high during tunnelling processes, it has to undergo recycling and is currently prohibited from entering natural water sources. Manometer Carbonic Acid In-flow from Oil Separation 1 m3 tunnel Container Suspended Lamell Flocculation Particles/pH Filters Chamber Flocculation pH Agent Artificial Sludge Dam Removal Figure 3.3. Schematic figure of the current water treatment plant for process water. The current water treatment plant visualized in Figure 3.3 is nonetheless performing efficiently and is not an issue since it keeps the previous mentioned parameters under control. The problem is the absence of biological treatment to reduce nitrogen compounds. 3.4 NITROGEN CONCENTRATIONS IN EFFLUENTS There is limited information about nitrogen removal and nitrogen concentrations in tunneling operations. However, the contaminated water i.e. the process water can be compared to effluents in mining activities which is also polluted by excessive nitrogen from explosives. There is more literature on mining, which will be used in order to identify a typical range of nitrogen concentrations in effluents. The Swedish mine company LKAB is used for reference values in terms of outgoing process water. The values have been collected from 2011-2013 at two sites, Kiruna and Vitåfors. The temperature is similar to that in tunnel water with a range of 0-19°C and 1-21°C in Kiruna and Vitåfors with a median value of 3.7 and 7.5 respectively. The range of ammonium nitrogen is 10 to 1 000 µg/l and the median value is 430 µg/l at the site in Kiruna. The range in Vitåfors for ammonium nitrogen is 40 to 2 600 µg/l and the median value is 964 µg/l. Thus, ammonium nitrogen is higher in Vitåfors. The range for nitrate nitrogen is 2 100 to 28 900 µg/l and 2 000 to 49 800 µg/l in Kiruna and Vitåfors respectively and the median for the two sites are 18 400 and 29 550 µg/l. The range for total nitrogen varies between 25 700 to 43 900 µg/l with a median value CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 20
Chalmers University of Technology
of 35 450 µg/l at the site in Kiruna. The same parameter in Vitåfors has a range of 30 300 to 69 300 µg/l and a median value of 54 300 µg/l (Suup, 2014). One research on biological removal of ammonia and nitrate in Ontario, Canada on mine water had an inlet range of nitrate concentration between of 51 000-405 000 µg/l (which converted to nitrate-nitrogen is 11 730-93 150 µg/l). This was considered to be in the typical range of inlet nitrate concentrations. The general ammonia and nitrate concentrations in mine effluents can vary quite considerably with 10 000-40 000 µg/l for total ammonia and 25 000-300 000 µg/l for nitrate in the effluent water (Koren, et al., 2000). These ranges converted to total ammonia-nitrogen and nitrate-nitrogen is 8 200 to 32 800 µg/l and 5 750-69 000 µg/l respectively. Poláková et al. (2013) used mine water with ammonia-nitrogen concentrations of 10 000-11 000 µg/l and achieved 86 percent removal efficiency. Ammonia is not monitored to the same extent as ammonium, nitrate and nitrite, though it is interesting to know in which range ammonia is usually present in mine- and tunnel water which can be used for comparison. Other ammonia levels in mine water vary in the range of 500-26 800 µg/l (410-21 976 µg/l ammonia-nitrogen) which is based on a project in Finland where cold temperatures were analyzed in relation to the bacterial community. The nitrate concentrations in this project varied between 7 200-52 500 µg/l (1 656-12 075 µg/l nitrate-nitrogen) (Karkman, et al., 2011). Another research on cold mine water was carried out in Finland in 2007 with water from one gold mine and one chromite mine. This research showed that fixed-bed biofilm reactors could be used to remove ammonium and nitrate in low temperatures. The ammonium-nitrogen concentrations of the two dewatering systems were 24 200 and 17 700 µg/l respectively and 17 500 and 4 300 µg/l for nitrate-nitrogen concentrations. However, ammonium rich water was presented as 50 000-100 000 µg/l why the ammonium concentrations in the two mines could be considered as moderate or low (Zaitsev, et al., 2008). According to Tilly et al. (2006), bilge water in tunnelling usually has a nitrogen concentration of 100 000 to 1 000 000 µg/l compared to untreated wastewater with 20 000 to 40 000 µg/l. Landfill leachate could also be used as reference values for the process water, however, this water has much higher nitrogen concentrations. The composition of a landfill leachate, Tveta, in Södertälje, Sweden, had the range of 1 100 to 200 000 µg/l ammonium nitrogen between 1994 and 2002. The total nitrogen concentrations during the same period had a range of 900 to 230 000 µg/l (Kietlinska, 2004). The explosives SSE and ANFO were used in two different tunnels in the project Botniabanan, performed by Skanska Sweden AB, in order to analyze the nitrogen contaminants from different explosives. The Björnböle tunnel used SSE and had an average nitrogen discharge in the process water of 170 000 µg/l but the levels varied from 25 000 to 275 000 µg/l. The total amount of SSE was 573 kg with a specific charge of 2.82 kg/m3 in the first analysis. The second analysis had a total charge of 620 kg SSE which gives a specific charge of 3.05 kg/m3. The Namntall tunnel used ANFO as the primary explosive and the average nitrogen discharge in the process water was 40 000 µg/l but it varied from around 10 000-50 000 µg/l. When the handling of ANFO was improved, i.e. less ANFO outside the boreholes, the average level decreased to 25 000 µg/l. The total charge was 290 kg of ANFO, DynoRex and Dynocord which gives a specific charge of 2.1 kg/m3. Thus, the total use of explosives is higher with SSE compared to ANFO. However, the nitrogen amount discharge using SSE in relation to charged SSE in kilograms is half of that using CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 21
Chalmers University of Technology
3.5 GENERAL DISCHARGE DEMANDS Many municipalities in Sweden have similar restrictions to the most common measured parameters in wastewater. Stockholm, Trollhättan and Malmö vicinity are some examples of areas using guidelines based on Table 3.3 from Svenskt Vatten. However, there could be local variances depending on external environmental factors. Table 3.3 presents parameters that, according to Svenskt Vatten, could affect the treatment processes and the water quality and are applied to industrial process water (Svenskt Vatten, 2012). Table 3.3. Discharge guidelines for some parameters for industrial process water (Svenskt Vatten, 2012) Parameter Unit Level pH - 6.5-10 Lead, Pb μg/l 50 Cadmium, Cd μg/l 0 Chromium, Cr μg/l 50 Nickel, Ni μg/l 50 Silver, Ag μg/l 50 Zinc, Zn mg/l 0.2 Ammonium mg/l 60 + Oil index 𝑁𝐻4 mg/l 5-50 The Swedish Food Administration has set out some guidelines for many parameters in drinking water. Table 3.4 presents the guidelines for when the water is unsuitable for drinking water and the guidelines for when the water is potable. Table 3.4. Guidelines from the Swedish Food Administration (SLV, 2001) Parameter Unit Limit unsuitable Limit potable Cadmium μg/l 5.0 Copper mg/l 2.0 0.2 Chromium μg/l 50 Nickel μg/l 20 Nitrate mg/l 50 20 Nitrite mg/l 0.50 0.10 Ammonium mg/l 0.50 3.5.1 MUNICIPAL VARIATIONS The municipality of Norrköping has provided guidelines for discharge of stormwater to recipients of different sensitivities. These guidelines apply for stormwater discharge CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 23
Chalmers University of Technology
and are presented in Table 3.5. However, guidelines for process water from different activities could have different and more restricted limits. Table 3.5. Discharge guidelines from Norrköping municipality (Dagvattengruppen, 2009). Discharge source Parameter Unit From operational Subarea to To recipient with activity recipient without protective values protective values Phosphorus, P mg/l 0.250 0.175 0.160 Nitrogen, N mg/l 3.5 2.5 2.0 Lead, Pb μg/l 15 10 8 Copper, Cu μg/l 40 30 18 Zinc, Zn μg/l 150 90 75 Cadmium, Cd μg/l 0.5 0.5 0.4 Chromium, Cr μg/l 25 15 10 Nickel, Ni μg/l 30 30 15 Suspended solids, SS mg/l 100 60 40 Oil index mg/l 1.0 0.7 0.4 The guidelines for discharge demands in Gothenburg are referred to the concentration in the discharge point, i.e. the effluents. These guidelines have been developed using several Swedish legislations and environmental goals and are seen in Table 3.6. Table 3.6. Guidelines for effluent concentration (Göteborgs Stad Miljöförvaltningen, 2013) Parameter Unit Concentration in effluents Chromium, Cr μg/l 15 Cadmium, Cd μg/l 0.4 Lead, Pb μg/l 14 Copper, Cu μg/l 10 Zinc, Zn μg/l 30 Nickel, Ni μg/l 40 Oil index mg/l 1.0 pH - 6-9 Total phosphorus, TP μg/l 50 Total nitrogen, TN mg/l 1.25 TOC mg/l 12 Suspended solids, SS mg/l 25 CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 24
Chalmers University of Technology
3.5.2 DEMANDS IN GERUMSTUNNELN Nitrogen concentrations are not always incorporated in the guidelines unless there is a sensitive recipient. There are mainly two surface water sources near Gerumstunneln which are Pulsebäcken and Gerumsälven. These two recipients are considered as sensitive since the environmental assessment by the Swedish Nature Centre showed presence of many different fish as well as beavers near Gerumsälven (Trafikverket & Ramböll, 2012). An internal analysis of the two recipients showed increased nitrogen concentrations but also increased metal concentration during the construction which is related to the construction work. The Swedish Transport Administration set out maximum concentration values for ammonium, nitrate and nitrite in Pulsebäcken and Gerumsälven in Tanumshede prior to the tunnel project start-up. Table 3.7 presents maximum concentration values for the three parameters which show that Gerumsälven has higher nitrogen limits due to a higher flow. Table 3.7. Acceptable parameter limits according to the STA (Trafikverket & Ramböll, 2012) Parameter Pulsebäcken Gerumsälven Ammonium <78 µg/l <312 µg/l Nitrate NH4−N <230 µg/l <1150 µg/l Nitrite NO3−N <15 µg/l <30 µg/l pH NO2−N 6-8 6-8 CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 25
Chalmers University of Technology
4 BIOLOGICAL NITROGEN REMOVAL TESTS The aim with tests in the lab is to try to get the biological treatment to work in terms of nitrification and denitrification of the process water. Nitrification and denitrification tests using water with temperatures below 10°C will not be analyzed, thus the water is at room temperature (19°). The reason for this is that the aim was to first see if it is possible to conduct a complete nitrification test at all at room temperature. The first experiment (Test 1) used only the process water and the nitrifying and denitrifying sludge while the second test (Test 2) used, besides water and sludge, sodium dihydrogen phosphate to compensate any potentially low phosphate concentrations. The third nitrification test (Test 3) increased the phosphate concentration as in Test 2 but the alkalinity was also increased. The reagents needed for the test were process water, sludge and compressed air. Process water was collected from two different days, 2014-03-31 and 2014-04-03. The water was stored in a cool place and then brought to Chalmers. A denitrification test was also carried out which used the process water from 2014-04-03 and a carbon source. 4.1 PREPARATIONS Before starting the nitrification process, some preparations were needed. The sludge concentration needed to be analyzed, the pH, the alkalinity and the nitrogen concentrations of the process water and sludge liquor had to be known in order to compare the results from the nitrification process, i.e. if there have been any activity during the test. The pH was measured to 6.5 in the process water from 2014-03-31 and 6.4 in the water from 2014-04-03. Process water from the two different sampling days was poured into three 50 ml test tubes each. These samples were frozen until they could be analyzed in the ion chromatograph which will give the concentrations of ammonium, nitrite and nitrate. The alkalinity was decided for the two different samples of process water from 2014- 03-31 and 2014-04-03 by the use of titration. 50 ml water each from the two days was used. Three drops of the indicator solution were added which turned the solution turquoise (at a pH of 5.4). The aim of this method (Standard methods, EN ISO 9963- 1:1995) is to titrate the turquoise solution until it becomes transparent. The titration solution used in this test was 0.02 M hydrochloric acid. The required volume of hydrochloric acid in the water to obtain a transparent color was 5.4 ml for the process water from 2014-03-31 and 5.3 ml for the water from 2014-04-03. When these volumes have been determined, the alkalinity can be calculated using Equation 4.1. The alkalinity for the two different water samples is calculated in Equation 4.2 and 4.3. (Eq 4.1) 𝑐(𝐻𝐶𝑙)×𝑉5×1000 𝐴 = 𝑉4 𝐴 = 𝐴𝑙𝑘𝑎𝑙𝑖𝑛𝑖𝑡𝑦,𝑚𝑀/𝑙 𝑐(𝐻𝐶𝑙) = 𝐶𝑜𝑛𝑐𝑒𝑛𝑡𝑟𝑎𝑡𝑖𝑜𝑛 𝑜𝑓 𝑡ℎ𝑒 ℎ𝑦𝑑𝑟𝑜𝑐ℎ𝑙𝑜𝑟𝑖𝑐 𝑎𝑐𝑖𝑑,𝑀/𝑙 𝑉4 = 𝑉𝑜𝑙𝑢𝑚𝑒 𝑜𝑓 𝑡ℎ𝑒 𝑤𝑎𝑡𝑒𝑟 𝑠𝑎𝑚𝑝𝑙𝑒,𝑚 𝑙 𝑉5 = 𝑉𝑜𝑙𝑢𝑚𝑒 𝑜𝑓 𝑡ℎ𝑒 ℎ𝑦𝑑𝑟𝑜𝑐ℎ𝑙𝑜𝑟𝑖𝑐 𝑎𝑐𝑖𝑑 𝑐𝑜𝑛𝑠𝑢𝑚𝑒𝑑 𝑏𝑦 𝑡ℎ 𝑒 𝑠𝑎𝑚𝑝𝑙𝑒 ,(𝑚E q𝑙 4.2) 0.02×5.4×1000 𝑚𝑀 𝐴(2014/03/31) = 50 = 2.16 𝑙 ×61 = 131.76 𝑚𝑔/𝑙 (Eq 4.3) 0.02×5.3×1000 𝑚𝑀 𝐴(2014/04/03) = 50 = 2.12 𝑙 ×61 = 129.32 𝑚𝑔/𝑙 CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 26
Chalmers University of Technology
The stoichiometry for nitrification and alkalinity is seen in Reaction 4.1 (Tchobanoglous, et al., 2004). For every mole of ammonium nitrogen to be transformed to nitrate nitrogen, two moles of HCO are required. 3 (Re 4.1) + − − However, for this test it is assumed that 2.5 mole of alkalinity is required per one 𝑁𝐻4 +2𝐻𝐶𝑂3 +2𝑂2 → 𝑁𝑂3 +2𝐶𝑂2+3𝐻2𝑂 mole ammonium nitrogen to make sure that the nitrification process is not limited by alkalinity. This relation was used at Ryaverket in Gothenburg and was used to see if the alkalinity was high enough for the ammonium nitrogen concentrations. PROCESS WATER 2014-03-31 (Eq 4.4) 𝐴 = 2.16 𝑚𝑀 𝑚𝑔 104 𝑁𝐻4 −𝑁 = 104 𝑙 => 14 = 7. 43 𝑚𝑀 (Eq 4.5) 𝐴Th𝑟𝑒e𝑞 r =eq 7u .i 4r 3ed × a 2lk .5al =ini 1ty 8 .5f 8or 𝑚 a 𝑀 concentration of 104 mg/l is 18.58 mM according to calculations in Equation 4.4 and 4.5, but the real alkalinity is only 2.16 mM. Thus, the alkalinity is too lo𝑁w𝐻 a4 n−d c𝑁ould inhibit the nitrification process. PROCESS WATER 2014-04-03 (Eq 4.6) 𝐴 = 2.12 𝑚𝑀 𝑚𝑔 111 𝑁𝐻4 −𝑁 = 111 𝑙 => 14 = 7. 93 𝑚𝑀 (Eq 4.7) 𝐴Th𝑟𝑒e𝑞 a =lka 7l .i 9n 3ity × w 2a .5s =cal 1c 9u .l 8a 3te 𝑚d 𝑀for this process water in Equation 4.6 which was 7.93 mM. The calculated required alkalinity in Equation 4.7 was 19.83 mM, thus the alkalinity in the water is too low in this sample which could limit the nitrification activity. SUPERNATANT 2014-05-02 The supernatant is the top layer liquid produced when the bigger particles of the sludge is allowed to settle. The alkalinity for the supernatant from May 2 was determined using the same method as for the process waters with a 50 ml sample and a 0.02 M acid. The required volume of the hydrochloric acid to make the sample go from turquoise to transparent was 5.8 ml. The alkalinity of the sludge is seen in Equation 4.8: (Eq 4.8) 0.02×5.8×1000 𝑚𝑀 𝐴4(.220 14/N05I/T0R2)I=FICAT50ION T=E2.S3T2 1𝑙 ×61 = 141.52 𝑚𝑔/𝑙 The first nitrification test was conducted on April 7 2014 where 1 500 ml samples of process water from the two days were mixed with 500 ml sludge using a magnetic stirrer. The solution was continuously aerated throughout the experiment. A total number of 26 samples, each 20 ml in size, were collected from the two solutions according to the schedule in Table 4.1. Each 20 ml sample was first centrifuged 4 000 RCF for 5 min. Thereafter the samples were filtered through 0.45 µm and stored in the refrigerator. These samples were later frozen for analysis due to maintenance problems of the ion chromatograph. The temperature was around 19°C in the two containers during the nitrification test. CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 27
Chalmers University of Technology
and with 31 samples, including two tubes of MQ water, the analysis was run for approximately 10 hours. However, there were certain preparations required before using the ion chromatograph. The 26 samples from the nitrification test were diluted 5 times with 2 ml of the sample and 8 ml MQ water. Three more samples were run in the chromatograph, the filtered samples of the process water and the sludge which gives a total of 29 samples. The first and last sample in the ion chromatograph is always filled with only MQ water which is used as a calibration to see that the instrument is working properly. 4.2.1 SLUDGE CONCENTRATION TEST 1 The sludge was collected from Hammargårdens treatment plant in Kungsbacka on April 7 (for Test 1 and 2) and May 2 (for Test 3 and the denitrification). This sludge consists of both nitrifying and denitrifying bacteria which are required for a nitrification and denitrification processes. The sludge concentration was approximately 4.6-5 g/m3 or 4 600-5 000 mg/l on April 7 and 2.3 mg/l on May 2 according to the instrument at site , however, the concentration was measured during the lab at Chalmers as well to obtain a more accurate concentration. First, the sludge was allowed to settle for a few minutes in order to remove most of the supernatant. 5 ml samples were collected of the sludge before weighing 1.6 µm filters which was required for the calculations of sludge concentrations. The 5 ml samples were then filtered through the 1.6 µm filters after which the filters were heated in the oven in 105°C for 2.5 hours for filter 1 and 2 (the minimum time for heating is 2 hours). The filters were cooled before being weighed a second time. The sludge sample from Test 2 was diluted with the process water, thus lower concentration. The filters were then heated in the oven again, but this time in 550°C for 30 min in order to measure the organic material content. Suspended solids could be calculated if the weight of the filter before and after heating in 105°C is known. The following equation is used: (Eq 4.9) 𝑋𝑑𝑟𝑦−𝑋𝑤𝑒𝑡 𝑣 𝑋𝑑𝑟𝑦 = 𝑡ℎ𝑒 𝑤𝑒𝑖𝑓𝑔ℎ𝑡 𝑜𝑓 𝑡ℎ𝑒 𝑓𝑖𝑙𝑡𝑒𝑟 𝑎𝑓𝑡𝑒𝑟 2 ℎ𝑜𝑢𝑟𝑠 ℎ 𝑒𝑎𝑡𝑖𝑛𝑔 𝑖𝑛 105° , l 𝑋𝑤𝑒𝑡 = 𝑡ℎ𝑒 𝑤𝑒𝑖𝑓𝑔ℎ𝑡 𝑜𝑓 𝑡ℎ𝑒 𝑓𝑖𝑙𝑡𝑒𝑟 𝑏𝑒𝑓𝑜𝑟𝑒 ℎ𝑒𝑎𝑡𝑖𝑛𝑔 The weight of the filters and the suspended solids are presented accordingly: 𝑣 = 𝑣𝑜𝑙𝑢𝑚𝑒 𝑜𝑓 𝑡ℎ𝑒 𝑠𝑎𝑚𝑝𝑙𝑒 Filter 1 • Before usage 0.0897 g • After heating in 105°C 0.1241 g • After heating in 550°C 0.0946 g Suspended solids 0.1241−0.0897 𝑔 6.88 Filter 2 0.005 = 6.88 𝑙 => 4 = 1.72 𝑔/𝑙 • Before usage 0.0888 g • After heating in 105°C 0.1249 g • After heating in 550°C 0.0941 g CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 29
Chalmers University of Technology
Suspended solids 0.1249−0.0888 𝑔 7.22 The nitrification rate 0s .0c 0a 5n no=w 7b.2e 2c 𝑙al=cu>late 4d =for1 .t8h1e 𝑔tw/𝑙o filters since the sludge concentrations is known. However, since this test failed to achieve a nitrification process, the rates will not be relevant. A 50 ml sample of the supernatant was also filtered through 1.6 µm on April 7 which was frozen for further analysis of the sludge content. The pH of the sludge, measured in the supernatant on April 7, was 6.84. The sludge age is not known but was approximated to be between 15-20 days according to the staff at Hammargården. 4.2.2 RESULTS TEST 1 The results from the ion chromatograph fluctuated and did not present values for all samples. The ion chromatograph turned out to give invalid results, this was determined after the instrument was run with some standards solutions. These samples were however kept and frozen until a new analysis was carried out when the instrument was repaired. A second analysis with the samples from April 7 was run on April 15 under correct conditions of the instrument. The concentration of phosphate was insignificant and below the detection limit, which could be the reason for why the biological treatment failed at site in Tanumshede. The theoretical sample 0 (start concentration after mixing process water with activated sludge suspension) is calculated to compare with the results from the ion chromatograph using Equation 4.10. (Eq 4.10) 1.5 ×𝑃 +0.5×𝑆 𝑆 𝑎𝑚𝑝𝑙𝑒 0𝑋 = 2 𝑆𝑎𝑚𝑝𝑙𝑒 0𝑋 = 𝑡ℎ𝑒 𝑓𝑖𝑟𝑠𝑡 𝑠𝑎𝑚𝑝𝑙𝑒 𝑖𝑛 𝑡ℎ𝑒 𝑛𝑖𝑡𝑟𝑖𝑓𝑖𝑐𝑎𝑡𝑖𝑜𝑛 𝑡𝑒𝑠𝑡 𝑃 = 𝑐𝑜𝑛𝑐𝑒𝑛𝑡𝑟𝑎𝑡𝑖𝑜𝑛 𝑖𝑛 𝑡ℎ𝑒 𝑝𝑟𝑜𝑐𝑒𝑠𝑠 𝑤𝑎𝑡𝑒𝑟,𝑚𝑔/𝑙 The theoretical samples for ammonium, nitrite and nitrate are calculated for the 𝑆 = 𝑐𝑜𝑛𝑐𝑒𝑛𝑡𝑟𝑎𝑡𝑖𝑜𝑛 𝑖𝑛 𝑡ℎ𝑒 𝑠𝑢𝑝𝑒𝑟𝑛𝑎𝑡𝑎𝑛𝑡,𝑚𝑔/𝑙 process waters from 2014-03-31. 1.5 ×104 +0.5×5 𝑆𝑎𝑚𝑝𝑙𝑒 01 (𝑁𝐻4 −𝑁) = = 79.3 𝑚𝑔/𝑙 2 1.5 ×5.6 +0.5×1.6 𝑆𝑎𝑚𝑝𝑙𝑒 01 (𝑁𝑂2 −𝑁) = = 5.4 𝑚𝑔/𝑙 2 1.5 ×346 +0.5×2.9 W𝑆𝑎h𝑚en𝑝 𝑙c𝑒o 0m1p (a𝑁ri𝑂ng3 −the𝑁 )th=eoretical values with the a=na2ly6z1e.d7 𝑚va𝑔lu/e𝑙s , 97.2, 5 and 305.6 2 mg/l respectively, it is clear that there have been some dilution faults. The theoretical first values of the samples using process water from 2014-04-03 are also calculated. 1.5 ×111 +0.5×5 𝑆𝑎𝑚𝑝𝑙𝑒 01 (𝑁𝐻4 −𝑁) = = 84.5 𝑚𝑔/𝑙 2 1.5 ×6 +0.5×1.6 𝑆𝑎𝑚𝑝𝑙𝑒 01 (𝑁𝑂2 −𝑁) = = 4.9 𝑚𝑔/𝑙 2 1.5 ×386 +0.5×2.9 𝑆𝑎𝑚𝑝𝑙𝑒 01 (𝑁𝑂3 −𝑁) = = 290.2 𝑚𝑔/𝑙 2 CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 30
Chalmers University of Technology
The values from the ion chromatograph show 71.7, 4.6 and 191.2 mg/l respectively which also indicate that there are some errors with the dilution. The results from the ion chromatograph were therefore normalized against sodium (anion) and chloride (cation) to get more accurate results. These compounds are expected to be constant during the experiment and by normalizing the data with these concentrations the errors in the measurements should be eliminated. Normalization using sodium had the best correlation and is presented in Figure 4.2. a) d) b) e) c) f) Figure 4.2. Results from Nitrification Test 1. Normalized values in figures a-c show how well sodium is correlating with the compounds. Figures d-f show the trend for ammonium, nitrite and nitrate. T he normalization was carried out by analyzing the ammonium, nitrite and nitrate concentrations in relation to sodium and chloride in order to consider the potential dilution errors. Thus the zero sample of the sodium concentration of the process water was used for samples when normalizing with sodium. The first (sample zero) was divided with the sodium concentration of the next sample X and then multiplied with n the ammonium concentration of that same sample, seen in Equation 4.11. The first sodium sample is then divided with the third sodium sample X and multiplied with n+1 the third ammonium sample. The same procedure is performed with nitrite and nitrate and with chloride as the normalizing agent. Sodium and ammonium is thereafter CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 31
Chalmers University of Technology
Nitrification Test 1 TOC 10 8 6 31-mar 4 03-apr 2 0 Time (hours) 0 5 10 15 20 25 Figure 4.4. Results TOC Nitrification Test 1 The concentrations of TN are presented in Figure 4.5. The concentration of TN should be on a similar level throughout a nitrification test since ammonium should be transformed into nitrate. However, in Test 1 the concentration of TN varied with a slight trend of increasing which could be due to disintegration of sludge flocs during the experiment. Nitrification Test 1 TN 65 55 45 31-mar 35 03-apr 25 Time (hours) 0 5 10 15 20 25 Figure 5.5. Results TN Nitrification test 1 4.3 NITRIFICATION TEST 2 This nitrification test was carried out on April 10 using the same method as Test 1. However in this test, 0.078 g sodium dihydrogen phosphate was added due to the low phosphate concentrations in Test 1 (which were below detection limit). This gives a phosphorus concentration of 10 mg/l and could help the nitrification since there required phosphate concentration could be too low. Samples were taken out according to the schedule in Table 4.2 and the pH was measured. All samples were frozen after Test 2 for further analysis. The temperature was around 16-17°C throughout the experiment. CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 33 )l/gm( NT )l/gm( COT
Chalmers University of Technology
Table 4.2. Sampling scheme for nitrification Test 2 Time Process water Process water 2014-03-31 2014-04-03 Min pH Sample pH Sample 0 6.9 0 6.84 0 15 - 1 - 1 30 - 2 - 2 45 - 3 - 3 60 8.31 4 8.25 4 90 - 5 - 5 140 - 6 - 6 170 - 7 - 7 220 8.31 8 8.25 8 1030 8.01 9 6.56 9 4.3.1 SLUDGE CONCENTRATION TEST 2 The sludge concentration in Test 2 was determined by the same method as for Test 1 with the only difference that the 5 ml sample was taken from the mixed solution in this test (process water and sludge). Filters 3 and 4 were heated in the 105°C oven around 14 hours (they were kept in the oven overnight) and for 30 min in the 550°C oven. Filter 3 • Before usage 0.0908 g • After heating in 105°C 0.1020 g • After heating in 550°C 0.0936 g Suspended solids 0.1020−0.0908 Filter 4 0.005 = 2.24 𝑔/𝑙 • Before usage 0.0878 g • After heating in 105°C 0.0962 g • After heating in 550°C 0.0885 g Suspended solids 0.0962−0.0878 4.3.2 RES0.U00L5TS T=E1S.6T8 2𝑔 /𝑙 The ion chromatograph was used on April 22 for the samples from the Nitrification Test 2. The theoretical sample 0 for ammonium, nitrite and nitrate are calculated for the process waters from 2014-03-31 using the same equation as in Section 4.2.2. The CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 34
Chalmers University of Technology
TOC is still low in concentration and varies throughout the test. TN seems to first decrease before it increases again, thus there is no complete nitrification in this test as already declared. Nitrification Test 2 TN 120 100 80 60 31-mar 40 03-apr 20 0 Time (hours) 0 5 10 15 20 Figure 4.7 Results TN Nitrification Test 2 4.4 NITRIFICATION TEST 3 Since neither Test 1 nor Test 2 resulted in any evident nitrification activity, another test was carried out, Test 3. It was suspected from the alkalinity tests that the alkalinity could be to low which was considered in this test. The proportion of the process water and sludge was altered in this test to see if that could have been a factor. The solution for the nitrification consisted of 1 000 ml process water and 1 000 ml sludge. The phosphate was, as in Test 2, increased by adding 0.078 g sodium dihydrogen phosphate in each container. The alkalinity was increased in this test by adding 1.34 g sodium hydrogen carbonate which was determined by looking at the ammonium concentration of the process water and the stoichiometry of alkalinity (estimated that one mole of ammonium nitrified consumes two moles of alkalinity). The following calculations were made in order to determine the required volume of sodium hydrogen carbonate to increase the alkalinity to meet the ammonium nitrogen concentrations: 𝑁𝑎𝐻𝐶𝑂3 • Process water March 31 𝑚𝑔 𝑁𝐻4−𝑁 104 52 𝑙 2• =Pro2ces=s w1a4ter= A3pr.7il 𝑚3 𝑚𝑜𝑙𝑒 3.7×2 = 7.4 𝑚𝑚𝑜𝑙 𝑚𝑔 𝑁𝐻4 −𝑁 111 55.5 𝑙 The test w=as run =for 22 hour=s a4n 𝑚d 𝑚15𝑜 𝑙m𝑒i n4 w×it2h =a 8to 𝑚tal𝑚 o𝑜f𝑙 12 samples per container. 2 2 14 Samples were collected at continuous intervals presented in Table 4.3 with the majority of the samples collected during the first 5 hours. CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 37 )l/gm( NT
Chalmers University of Technology
Table 4.3. Sampling scheme nitrification Test 3 Process water Process water Time 2014-03-31 2014-04-03 Min pH Sample pH Sample 0 7.83 0 7.82 0 15 - 1 - 1 30 - 2 - 2 45 - 3 - 3 60 8.29 4 8.16 4 90 - 5 - 5 120 - 6 - 6 150 - 7 - 7 180 8.26 8 7.98 8 240 - 9 - 9 300 8.14 10 7.85 10 1335 7.79 11 7.97 11 4.4.1 SLUDGE CONCENTRATION TEST 3 The sludge concentration was determined with the same method that was used in Test 1. The 5 ml sample that was filtered was taken from the sludge solution after it had been thickened i.e. much of the supernatant had been removed. Filters 5 and 6 were heated for 2.5 hours in 105°C and 30 min in 550°C. Filter 5 • Before usage 0.0949 g • After heating in 105°C 0.1505 g • After heating in 550°C 0.0885 g Suspended solids: 0.1505−0.0949 𝑔 11.12 Filter 6 0.005 = 11.12 𝑙 => 4 = 2.78 𝑔/𝑙 • Before usage 0.0926 g • After heating in 105°C 0.1424 g • After heating in 550°C 0.1003 g Suspended solids: 0.1424−0.0926 𝑔 9.96 The nitrification ra te c 0a .n 00 b 5e cal=cul9a.t9e6d 𝑙us=in>g th 4e s=lu2d.g4e9 c 𝑔o/n𝑙centration in Section 4.2.1 and Equation 4.12. However, since the nitrification only worked from ammonium to nitrite, which is the only nitrification rate that is possible to obtain. These calculations CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 38
Chalmers University of Technology
undetected during the analysis, the additive of sodium hydrogen carbonate could be incorrect in order to achieve the wanted alkalinity. Test 3 was also analyzed for TOC and TN. The result for TOC is presented in Figure 4.10 for the two different types of process water. The figure shows an increasing curve of TOC concentration, though; there are some negative values for the process water from April 3. This could be a result of the concentration being very close to detection limit, thus limits that the machine is unable to discover. Therefore the machine could have resulted with some negative values in these measurements. Nitrification Test 3 TOC 4 3 2 1 31-mar 0 03-apr 0 5 10 15 20 25 -1 Time (hours) -2 Figure 4.10. Results TOC nitrification test 3 Figure 4.11 shows the trend for TN for both process waters. The results for the two days are very similar and show that TN is staying at the same level, which indicates that some nitrification might be occurring since ammonium turned into nitrite. Nitrification Test 3 TN 50 40 30 20 31-mar 10 03-apr 0 Time (hours) 0 5 10 15 20 25 Figure 4.11. Results TN Nitrification Test 3 4.5 DENITRIFICATION TEST A denitrification test was also carried out in the lab at Chalmers University on May 5 using sludge from May 2 collected at Hammargården wastewater treatment plant. The process water from April 3 was the only type used in this test due to the similarities in water compositions between March 31 and April 3. April 3 had slightly higher concentrations of nitrate why this sample was chosen. The sludge was first thickened by removing some of the supernatant. It was also aerated for 30 minutes before used CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 41 )l/gm( COT )l/gm( NT
Chalmers University of Technology
in the experiment since it had been stored cool for a few days. 1 100 ml of water and 900 ml sludge were first mixed by a magnetic stirrer. Nitrogen gas was then added for 30 min before the carbon source was added. The carbon source was acetate and the required amount was calculated to 2.86 g/2 l using the stoichiometry fo−r acetate and the concentration of nitrate which was determined to 386 mg𝐶/l𝐻 i3 n𝐶 𝑂th2 e nitrification test. The experiment was run for 3 hours (based on other standard denitrification tests) during regularly sampling, which is seen in Table 4.5. Sample 0 A is when acetate was added but due to measurement errors it was only 1 900 ml in total why another 100 ml of process water was added when sample 0 was taken. The B samples were centrifuged for 5 min as in the nitrification tests and then filtered directly through a 0.45 µm filter. The nitrogen samples were diluted 5 times using 2 ml filtered samples and 8 ml MQ water. These samples were then run in the ion chromatograph. Table 4.5. Sampling schem e for denitrification test Process water Time 2014-04-03 Min pH Sample 0 - 0 A 8 - 0 B 15 7.97 1 30 - 2 45 - 3 60 - 4 90 8.29 5 120 - 6 150 - 7 180 8.47 8 4.5.1 SLUDGE CONCENTRATION DENITRIFICATION The sludge concentration in the denitrification test was also determined by using the same method as described in Test 1, Section 4.2.1. Filters 7 and 8 were heated for 2.5 hours in 105°C and 30 min in 550°C. Filter 7 • Before usage 0.0954 g • After heating in 105°C 0.1431 g • After heating in 550°C 0.1017 g Suspended solids: 0.1431−0.0954 𝑔 9.54 Filter 8 0.005 = 9.54 𝑙 => 4 = 2.39 𝑔/𝑙 • Before usage 0.0934 g CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 42
Chalmers University of Technology
• After heating in 105°C 0.1406 g • After heating in 550°C 0.0999 g Suspended solids: 0.1406−0.0934 𝑔 9.44 4.5.2 RES0U.00L5TS D=E9N.4IT4R𝑙 =IF>ICA4 T=IO2N.3 6T 𝑔E/S𝑙T Figure 4.12 presents the results from the denitrification test. The denitrification process should give decrease in nitrate before transformed to nitrite which in turn should be transformed to nitrogen gas. The results show increasing nitrite and decreasing nitrate which indicates that the denitrification process was working to some extent. However, the test was only conducted during 3 hours at which the nitrate concentration was measured to around 110 mg/l, which is still a high concentration since the goal was to decrease nitrate close to zero mg/l. Thus the nitrate concentration was decreased around 20 mg/l while nitrite was increased with 40 mg/l. This shows that the denitrification was not fully complete since not all nitrate was transformed into nitrite. The pipette was working when the samples were diluted which would make the results valid. But the results were normalized as in Test 3 with sodium and chloride. a) b) c) Figure 4.12. Results denitrification test CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 43
Chalmers University of Technology
5 DISCUSSION The main purpose of this master thesis was to conduct nitrification and denitrification tests in the lab with process water from Gerumstunneln. The concentrations of ammonium, nitrite and nitrate were measured in the first test and the other tests were based on these concentrations. Though, when the process water was analyzed in the third nitrification test presented in Section 4.4.2, the nitrogen concentrations had increased. This could have affected the results in the test, in particular the calculations used to increase alkalinity in Test 3 since it was based on the nitrate concentrations from Test 1. 5.1 NITRIFICATION TESTS The results showed a small decrease in ammonium (which should occur), a very small increase in nitrite and finally fluctuating values for nitrate. This shows that there was no complete nitrification, nor a very good result for the start of the nitrification process. It was suspected that the phosphate concentration was too low in the process water why it was decided to increase the concentration in the next nitrification test. It was also believed that the dilution had not been performed with great accuracy due to the pipette used. Thus, the results were normalized using sodium (cation) and chloride (anion) and the theoretical first value was calculated. The theoretical values were not consistent with the measured values why the conclusion was that the dilution had been compromised during the use of the pipette, where the volume in the test tubes varied from around 9.5-10.7 ml, in which there should have been exactly 10 ml. The correlation between sodium and ammonium, nitrite and nitrate were very good. The normalization with chloride also showed good correlation but sodium proved to be the best one why the diagrams and results are based on sodium. Since the results showed no clear sign of nitrification, a new test needed to be carried out. The concentration of TN varied during Test 1, which should have been at the same level throughout the entire test if the nitrification had been completed since ammonium is transformed into nitrate, thus the amount of TN should still be the same. But with the result presented in Section 4.2.2, it shows that the process was not working properly. The reason for both increases and decreases of TN during the test could be the lack of phosphate, low alkalinity and possibly lack of aeration during certain periods due to irregularly pressure in the flow. It could also have been simultaneous denitrification due to the irregular flow if the aeration was insufficient. The second nitrification test used the same process water as in Test 1. Since the results in Test 1 did not show nitrification activity as hoped, it was decided to increase the phosphate concentration to see if it would help the process. These samples were diluted at the same time as Test 1 (the samples had been frozen) which gave Test 2 the same dilution errors as Test 1 since the same pipette was used. The results were therefore normalized with sodium and chloride to see if any of the two had better correlation. Sodium proved to have better correlation again and these values were used when analyzing the curves and ammonium, nitrite and nitrate were plotted against the time. Ammonium should decrease during nitrification, and since phosphate had increased the hope was to achieve better results. However, the trend for ammonium or nitrite was not very distinctive. This shows that nitrification was not working in this test either, despite increase of phosphate. The pH was measured in the beginning of the test to around 6.9 for both process water solutions, but the process water from March 31 increased during the test to 8.3 after 220 min but stopped at 8 CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 45
Chalmers University of Technology
after the entire test. The process water from April 3 also reached around 8.3 during the test but when the pH was measured after 17 hours, it had decreased to 6.6. The pH should be between 6.5 and 8.0 for nitrification to occur, which the pH was during the test, thus the pH is most likely not the parameter causing nitrification failure. The concentration of TN during Test 2 fluctuated, as in Test 1, which further proves that this test had an incomplete nitrification. The third nitrification test used sludge collected the same day from the wastewater treatment plant. From the previous two tests, and since something in the project was decreasing the alkalinity in the recipients, it was suspected that the alkalinity was too low. The alkalinity was determined to 2.16 mM and 2.12 mM for the process waters which were below the required alkalinity for the high concentration of ammonium in the waters. The alkalinity was therefore increased by adding sodium hydrogen carbonate. The phosphate concentrations were increased to the same levels as in Test 2. The pH was monitored and it was around 7.8 in the beginning of the test for both process waters. The process water from March 31 increased to 8.3 but decreased to 7.8 at the end of the test. The process water from April 3 also increased in the middle of the test to 8.2 and then decreased to 8 at the end. The pH has been in the range of what is preferable for nitrification (6.5-8.0) which should not have impacted the nitrifying activity. If pH increases in the solution, thus becoming more basic, ammonium can transform into free ammonia. However, since pH has been kept at a stable level, this should not be the reason for nitrification failure. The pipette had undergone maintenance and gave good results without normalizing with sodium and chloride. Nonetheless, the final step in the nitrification process i.e. nitrite to nitrate, did not work in this test either. The nitrate concentrations were kept at a similar level, thus an incomplete nitrification. The ammonium levels were close to zero mg/l at the end of the nitrification test but if the test would have been kept running longer, it is a possibility that all ammonium would have transformed into nitrite which maybe would have started to transform into nitrate. The concentrations of TN during Test 3 remained at the same levels which indicate that some nitrification is occurring since the amount of nitrogen compounds is similar during the entire test with the exception of a small increase at the end of the test. Thus, nitrification was occurring at the beginning of the test but was inhibited between the nitrite-nitrate transformations. 5.2 DENITRIFICATION TEST There were only one denitrification test performed in this study. Denitrification requires a carbon source and acetate was used for which the amount was determined based on the nitrate concentrations from Test 1. The phosphate and the alkalinity concentrations were not increased in this test in order to observe the performance of only process water, sludge and acetate. The aim with denitrification is to change nitrate into nitrite which turns into nitrogen gas. The dilution for this test was also done accurately with the pipette and should not need normalization. The results show that nitrate is decreasing while nitrite is increasing. Ammonium is also decreasing but with some fluctuating values. The results are quite good for the denitrification, even though not all ammonium is removed. The test was only run for 3 hours, which decreased nitrate from 130 mg/l to 110 mg/l (Δ 20 mg/l) while nitrite increased from 10 mg/l to 50 mg/l (Δ 40 mg/l). If the test would have been run for longer time, e.g. around 20 hours as in the nitrification tests, the nitrate could have been further decreased and nitrite further increased. Denitrification aims to decrease nitrogen levels and the concentration of TN shows a decrease during the test. This proves that some of the nitrate and nitrite are transformed into nitrogen gas which has left the CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 46
Chalmers University of Technology
solution, thus denitrification is partially in progress. However, when assessing the results it is seen in Figure 4.12 that all nitrate is not transformed into nitrite and all nitrite is not transformed into nitrogen gas. This test had other sludge-process water proportions compared to the nitrification tests, i.e. more sludge was used in relation to the volume of water. This could have contributed to better results and it would have been interesting to run a test with 50/50 proportions for all three nitrification tests as well to see if the results would be different. If phosphate and alkalinity concentrations would have been increased in this test it might have resulted in an even better outcome since it already showed some denitrifying activity. 5.3 SLUDGE The sludge used was collected from a treatment plant in Kungsbacka which had both nitrifying and denitrifying bacteria. Due to logistics, the sludge was only used fresh i.e. the day it was collected from the plant, for Test 1 and Test 3. For Test 2 and the denitrification test, the sludge had been kept cooled in containers for maximum 3 days. This could affect the processes even though the sludge was aerated before usage. It would have been preferable and probably better for the tests to have new daily sludge for all tests, but due to logistic difficulties of collecting the sludge, the same sludge was used in more than one test. However, another aspect of failed nitrogen removal could be related to the sludge in terms of acclimation to the environments for the bacteria. The bacteria in the sludge are not used to process water from a tunnel that was used in this study, i.e. those bacteria are used to regular wastewater. Bacteria sometimes need time to acclimate to new environments and time for acclimatization was not provided for the bacteria prior the tests. This could be a factor that should be assessed further, whether or not the bacteria in the sludge need to be familiar with the process water before running a nitrification and denitrification test. For further analysis, it would be interesting to run nitrification tests using the same process water but different sludge from different wastewater treatment plants to see if the processes differ or if the problem with nitrification only relates to the composition of the water. The sludge age could also be a factor inhibiting nitrification. The exact sludge age was never known, but an approximated guess at the wastewater treatment plant was 15-20 days. It would have been interesting to use sludge with higher age since sludge older than 20 days has proved to help the nitrification process in cold temperatures. 5.4 SUMMARY No tests in this analysis showed any distinct results on nitrification. They were all incomplete and the nitrification process only went from ammonium to nitrite, but it failed to change into nitrate which was the common factor for all three nitrification tests. Thus, the second step in the nitrification test is causing the failure and the reason for this is unclear. The denitrification process was quite successful since nitrate decreased, nitrite increased and TN decreased. However, since the total time was not sufficient enough to remove larger quantities of nitrogen, the test should have been performed much longer. The reason for why nitrification failed is still undetermined. Some thoughts and assumptions have been developed along the analysis. First, the concentrations of nitrogen in the process water are much higher than in other activities, such as in mining industries or in regular wastewater, which increases the pressure on the biological treatment. Some metals have inhibiting effects on nitrification such as zinc, copper and nickel, and some of these metals had increased in concentration in both Pulsebäcken and Gerumsälven during the tunnel project. The CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 47
Chalmers University of Technology
metal increase comes most probably from the process water in the tunnel. The impacts from metals have not been taken into account during the lab. This could therefore be a reason, in combination with other factors such as phosphate and alkalinity, of the nitrification failure. It is a possibility that the phosphate concentrations should have been increased even further in Test 2 and Test 3 in order to achieve a complete nitrification since the same concentration was used for the two. Same connection applies for alkalinity which could have been increased even more in Test 3. Other inhibiting parameters are pH and nitrite itself. However, pH has been quite stable during the tests and has almost at all times kept within the range of 6.5-8.0 and should not be the primary affecting parameter. The concentration of nitrite could also inhibit the process if it is too high but nitrite was below the critical levels. Another problem could also be a shortage of micronutrients that are necessary for microbial growth. The problem for projects such as Gerumstunneln is the cold water in the biological treatment which inhibits the nitrification process in terms of decreasing the nitrification rate. However, all tests in the lab were performed in room temperature and the nitrification process was still not completed. This indicates that the low temperature is not the major issue, even though it could be a problem as well. It was assumed prior to this study that nitrification would be possible in at least room temperature and that the problem would be to manage the cold water. The analysis in the lab does however present other complex results which brings more questions than just those about low temperatures. There is no clear answer to why the nitrification tests were uncompleted, but it is not only cold water that contributes to the failure, the proportions of sludge and process water and the concentrations of metals could also have been a strong factor to the results. CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 48
Chalmers University of Technology
6 CONCLUSION The conclusions of this master thesis are stated in this section. • The first nitrification test resulted in an incomplete nitrification which was assumed to be correlated with low phosphate concentrations in the process water. • The second nitrification test also showed incomplete nitrification despite a phosphate concentration increase. The phosphate increase could however have been too low in order to achieve a complete nitrification. • The third nitrification test showed the best results from the nitrification tests assumed to be correlated with an alkalinity increase in combination with an increase of the phosphate concentrations. Hence, the ammonium-nitrite transformation was working for all three tests but the second step of transforming nitrite to nitrate failed. This is most likely due to the phosphate and alkalinity concentrations, but also to the metal concentrations of the process water which was not assessed. • The denitrification test showed partial denitrification which could be related to a proportion alteration between process water and sludge compared to the nitrification tests and that denitrification is less sensitive. The issue prior this study was the low water temperatures in tunnels and in the biological treatment. All tests in the lab were performed in room temperature and the nitrification process still failed. Thus, the low water temperature at site is not the major concern. Other inhibiting parameters on nitrification are metals where the assessment on the recipients showed increases in several metals that could affect the nitrification process. The pH could also affect nitrification, though it should be between 6.5-8.0 which it was for most part of the tests. The proportions in the solutions of sludge and process water was 500/1 500 ml in the nitrification tests and 900/1 100 ml in denitrification test, which also had the best results. This could be a factor that affects the nitrification process in combination with increased metal concentrations, low alkalinity and low phosphate concentrations. Further studies are required to provide substantial answers and a solution to why the process water from tunnels is difficult to treat using biological treatment of nitrification and denitrification. Water is complex and cannot always be analyzed in a universal perspective, the local conditions differ and a tunnel in other parts of the world could have other water compositions and surrounding environments. Hence, further lab analysis on process water with high nitrogen content is crucial which should focus on phosphate and metal concentration, sludge age and the proportions of sludge and process water since those parameters were not thoroughly analyzed in this study. CHALMERS, Civil and Environmental Engineering, Master’s Thesis 2014: 49
Chalmers University of Technology
ii Concept development of field analysis equipment for mining and exploration application Master’s thesis in the Master Degree Programme Product Development VASUPOL KUNAVUTI Department of Product and Production Development Chalmers University of Technology Abstract The problems with the exploration process within the mining industry are high investment and high risks because it takes a very long time and a lot of money to perform the analysis of the rock sample. With the developed equipment, the process will be shortened, thus benefitting the mining and exploration companies who are the potential customers. The aim of this thesis is to create a feasible and production ready concept of equipment for performing field analysis of rock samples in mining and exploration application. During the course of this thesis work, a product development process has been used. This process involves identifying requirements specifications, functional analysis, concept generation, concept evaluation and concept refinement. The new product is a newly developed instrument that was improved from the first prototype. Mainly, the new feature of the machine is that it can scan multiple samples in a standard core box. The design emphasis of the instrument was on operability, usability, safety, transportability, environment and compact ability. The aim of the final design was to arrange the components in order to save space and weight, while maintaining the functionality and quality of the analysis. The three-dimensional CAD model prototype was created by using a computer-aided design program. The design was based on the final concept. The special focuses have been on the new feature of the instrument such as the loading and unloading of the core box. It is important that the final design is feasible for production, as the aim is to turn the concept into a fully functioning product. The result of the thesis provides the solutions that analyses multiple samples in a standard core box. The instrument can withstand the rough and tough environment, and the exploration process can be performed faster and more accurately. Keywords: Mining, Exploration, rock analysis, core sample, core box
Chalmers University of Technology
Chapter 1: Introduction 1 1. Introduction This chapter introduces the thesis which includes the proposal, objective and scope. In addition, the current situation of the market within the exploration and mining industry and the motivation of this thesis are also discussed. 1.1 Background The mining and exploration industry is growing every year. In Sweden alone, a lot of new projects are moving towards production. In the upcoming year, the newly established mines are scheduled to start production and existing mines are expanding their production. For example, Dannemora iron ore mine north of Stockholm is planning to start production in 2012. The Northland resource in the north outside Pajala, on the border to Finland, is scheduled to expand their production to five million ton in 2013. Figure 1.1 (Value of exploration in Sweden 1982-2010) shows that in 2010 investments in exploration rose to 675 million SEK. The number of exploration permits rose by 40% in the first quarter of 2011 and 192 permits were granted up until September 2011, while another 50 are being processed at the Mine Inspectorate in Luleå. 323 Exploration permits have been granted extensions and a further 95 decisions are pending. (Tomas From, 2011) Figure 1.1: Value of exploration in Sweden 1982-2010 (million SEK, current price) (Tomas From, 2011) In order to find the economically feasible mineral resources, the exploration drilling process is associated with high investment and major risk. Currently, mining and exploration companies use a portable XRF analyzer and geologist’s judgment to select the core samples from several drill holes reaching hundreds of meters below the ground and send those samples to laboratories for accurate analysis. Figure 1.2 illustrates the exploration process. This process takes a lot of time and is very costly. In view of the information mentioned above, it is a great opportunity for the company to develop an instrument that can analyze rock core samples right at the mine and exploration site. With this new instrument, the process will be shortened since, the instrument can produce the scanned result on-site more accurate and repeatable than the delivered results by the portable XRF analyzer. The samples no longer need to be selected because the developed machine can analyze the entire sample from the drill hole. Exploration Sample Sample Investment Pre-studies drilling selection analysis decision Figure1.2: The exploration process
Chalmers University of Technology
2 Chapter 1: Introduction The company decided to produce a single core scanner to reduce the complexity of the prototype. The next product will be an improvement on the design and function of the first prototype and will be produced in the future, based on the final concepts of this thesis. 1.3 Purpose The aim of this thesis is to generate feasible concepts of X-ray equipment based on the first prototype for performing field analysis of rock samples in mining and exploration applications. The developed equipment will be capable of analyzing multiple samples in a standard core box. 1.4 Objective The objective is to develop and redesign the current prototype equipment with emphasis on the following aspects:  Operability, the developed equipment should be able to analyze more rock samples than the existing prototype.  Usability, the machine should be easy to use according to ergonomics.  Safety, it should be safe to operate.  Transportation, the instrument should be easy to transport.  Environment, the equipment should be able to operate in the rough and tough environment.  Design, the instrument should be relatively small and light weight. The intended outcomes of the thesis are the concepts' solutions for the following functions: loading and unloading the core box, movement of X-ray components and core box, transport solution, maintenance, environment, components layout, safety and outer design. 1.5 Scope Within its parameters, the thesis focuses on developing a concept of field analysis equipment. The thesis does not include programming of PLC or control software, operating software, user interface software and no physical product. This thesis does not go deep into the subsystem. However, it does make suggestions on specific components’ layout and requirements. The developed equipment will be presented with CAD models. The thesis is planned to take 6 months to deliver the expected outcome. Following areas will be developed during this thesis:  Construction of the list of requirements specification  Generate, evaluate and select the concepts of loading and unloading sample, movement of samples and components, maintenance and transportation  Designing the layout of the instrument  Designing the cover  Create a CAD model and drawing
Chalmers University of Technology
4 Chapter 2: Thesis Process some part of the report was not updated and needs to be improved. The next step was to study the sub systems and components of the existing prototype and create component structures to show the whole system of the machine. The manual of the sub components helps identify the requirements of the sub system. A process flow model was made to illustrate the input process and output of the existing prototype. 2.3 Development phase The Development phase begins with identifying the function and problem. The function mapping was done to show interference between the functions and to find the most effective and affected functions. The next step is to generate the concepts. The ideas were established and sketched to visualize the instrument. A brainstorming session with the company’s representatives was set up to help develop the idea and to discuss alternative solutions. The concept table was used to present ideas and solutions of the whole system. The concepts were narrowed down by using evaluation methods such as Elimination matrices, the evaluations based on Go/No-Go screening and the Kesselring matrices. The Elimination matrices helped to eliminate unfeasible ideas. The evaluations based on Go/No-Go screening were used to screen the concepts. Finally the Kesselring matrices were used to select the promising ideas by applying rated criteria that were selected from the requirements after the discussion with the company’s representatives. The outcome concept is the result from using evaluation methods. It contains the solution of each function that suited the product. It is however possible that during the product development process the requirements might have changed or some solution might not be able to be applied to the equipment. The outcome concepts need to be refined in the next phase. 2.4 Refinement and deliverables phase In this phase, the outcome ideas were refined and finalized by assuring that the idea solutions are feasible and working well together. After establishing the final concept it is important to create a prototype of the product. The prototype is a visual representation of the concept. A CAD model was created by using the Autodesk inventor program. It is the best way to design the machine because if the concept is not feasible, or too complex to be accomplished in practice, the change of design can be made within the program. Finally, the report has been finished and presented.
Chalmers University of Technology
Chapter 3: Theoretical Framework 5 3. Theoretical Framework The method and tool that were used in the thesis including functional analysis, relations diagram, concept table, concept evaluation, failure mode and effects analysis and prototyping will be described in this chapter. 3.1 Functional analysis Functional analysis is the analysis of the system in terms of its purpose. It involves the description of the main function and sub-functions of the system. Two functional analysis methods were applied in the thesis, namely, Process flow model and Function means tree. 3.1.1 Process flow model The main purpose of a process flow model is to decompose the primary function into sub- functions. The flows regarding material, energy and information between the sub-functions are described. The method focuses on input operands, output operands and transforming function. Figure 3.1 illustrates the process flow model. This method was used in the thesis to define the flow of the analyzing process of the developed product in order to improve the desired function and design the new function. (Ullman, 1944) Input operand Function Output operand Figure 3.1: Process flow model 3.1.2 Function-means tree Function mean tree is an approach to capturing the functional requirements and decompose them in a hierarchical structure. It decomposes the sub-functions and sub-solutions, starting from the primary function. The purpose of this method is to define the interaction between the functional domain and physical domains. The tool consists of two types of nodes; functions - what needs to be done and means - how those can be done. This approach can also be a tool for idea generation. Figure 3.2 illustrates a function-means tree. Function-means tree was used in the thesis to provide an understanding of the relationship between functions and the components. (Ullman, 1944) Function Means Figure 3.2: Function-means tree 3.2 Relations diagram The relations diagram also known as interrelationship diagram is used to identify cause and affect relationships between different issues. The relation diagram is created by identifying factors that are involved in the problem; draw an arrow from the greater influence (the cause)
Chalmers University of Technology
6 Chapter 3: Theoretical Framework to the lesser influence (the one influenced). The factor with the most outgoing arrows will be the driver or causal factor and the factor with the most incoming arrows will be an outcome or result. This tool was used in the thesis to organize the functions of the instrument. It identifies causal factors and help to decide which function to focus on. (Charantimath, 2009) 3.3 Concept table The use of the concept table (see Table 3.1) was developed for this specific thesis. The concept table was used for two purposes, to collect the overall solutions in one table and to show the promising solutions that were evaluated from various methods. The alternative solutions were indicated by color code. The different colors mean that the solutions were excluded by different evaluation methods. The total amounts of possible solutions are massive and the functions of the developed product have unique characteristic, some can affect the complexity of the overall system, and some have individual properties, which can be influenced from other functions. To simplify the process, the sub-solutions have been evaluated and selected from each function before starting to combine the solution. The potential performance of a solution combination was synthesized to support the decision. Table 3.1: Concept table Solution Function/ 1 2 3 … m Sub-system F 1 S1.1 S1.2 S1.3 … S1.m F 2 S2.1 S2.2 S2.3 … S2.m F 3 S3.1 S3.2 S3.3 … S3.m : : : : : : F n Sm.1 Sn.2 Sn.3 … Sn.m 3.4 Concept Evaluation Concept evaluation is the process of evaluating and selecting the potential concepts from the idea generation stage. The aim is to narrow down the amount of concepts and decide which concepts have the highest possibility for becoming an actual product. The concept evaluation methods were used in the thesis, including the Elimination matrix, the evaluation based on Go/No-Go screening and the Kesselring matrix. 3.4.1 Elimination matrix The Elimination matrix is a filter for the concepts. The evaluation is based on feasibility judgment. An idea will be excluded if it fails to fulfill one of the following criteria: solving the main problem, fulfilling the demands of the requirement list, realizable in principle, reasonably cost effective and safe. Each solution will be judged against the criteria. The result can be: Yes (+), No (-), Lack of information (?), or Check requirement list (!). And the decision for each solution can be: Pursue solution (+), Eliminate solution (-), Collect information (?), or Check requirement list for changes (!). In the thesis, this method was used to eliminate obviously disqualified solutions. It is a quick step before concept screening and concept scoring. Table 3.2 shows the Elimination matrix. (Pahl, Wallace, & Blessing, 2007)
Chalmers University of Technology
Chapter 3: Theoretical Framework 7 Table 3.2 Elimination matrix Selection Criteria Solution Remarks Decision A B C D E S1 S2 : Sn 3.4.2 Evaluation based on Go/No-Go screening After that the concepts have passed the Elimination matrix, they will be evaluated against the criteria defined by the requirements specifications. The questions that generate from the requirements need to be answered with either Go, or No-Go. This method was modified to make it more suitable for the thesis. Instead of using the matrix, the evaluation was carried out in the meeting between the stakeholders. The discussion stemmed from which requirements should be selected to create the criteria preceding the screening process. (Ullman, 1944) 3.4.3 Kesselring matrix A Kesselring matrix as shown in Table 3.3 is a concept scoring method. It is used to analyze potential concepts and finally selecting the final concept. The processes of the Kesselring approach are:  Choose the criteria based on requirements.  Determine the weight factors.  Assign the grades for fulfillment of the selection criteria.  Calculate the concept scores. The final concept is selected from the best total score. The Kesselring matrix was applied in the thesis to evaluate and select the final solution of an individual function. The selection criteria and weight factors were carefully considered by the stakeholders. Each solution was assigned a numerical value corresponding to the level of fulfillment of the selection criteria. (Ulrich & Eppinger, 2008) Table 3.3 Kesselring matrix Evaluation Weight Concept 1 Concept 2 Concept m criteria (W) Value WV Value WV Value WV 1 2 m (V ) (V ) (V ) 1 2 m A W V W V V W V V W V 1 1A 1 1A 2A 1 2A mA 1 mA B W V W V V W V V W V 2 1B 2 1B 2B 2 2B mB 2 mB C W V W V V W V V W V 3 1C 3 1C 2C 3 2C mC 3 mC : : : : : : : : n W V W V V W V V W V n 1n n 1n 2n n 2n mn n mn ∑W ∑ WV ∑ WV ∑ WV 1 2 m 3.5 Failure mode and effects analysis The Failure mode and effects analysis (FMEA) is a technique for identifying failure potential and its consequences and the methods to reduce the chance of failure occurring. FMEA is used
Chalmers University of Technology
8 Chapter 3: Theoretical Framework as a design evaluation tool and as a design aide. A failure can be a mechanical failure, based on any chance that the component, assembly, or system fails to perform its intended function. Table 3.4 illustrates the FMEA worksheet. It contains four indicators - Potential failure, Potential effects of failure, Potential cause and Current controls. Also there are three ratings - Severity, Occurrence and Detection. (Otto & Wood, 2001) Severity relates to the seriousness of a potential effect of failure on a scale from 1 – 10. 1 implies no effect and 10, implies a very high and hazardous effect. The failure may result from the product becoming inoperative or unsafe and could result in possible injury. Occurrence is the probability of failure based on potential cause(s), over the lifespan of the product. This is indicated on a scale of 1 - 10, where 1 implies failure unlikely and 10 implies failure is almost inevitable. Detection relates to the possibility that current controls will detect the potential failure before sending the product to the customer. Finally, a Risk Priority Number (RPN) is calculated from these values. (Otto & Wood, 2001) Table 3.4: FMEA Worksheet Potential Potential Risk priority Severity Potential Occurrence Current Detection Function Failure Effects of number (S) Cause(s) (O) controls (D) mode Failure (RPN) 3.6 Prototyping Prototyping is the process of developing an approximation of the product. It provides an efficient and effective method to modify and improve the product through the testing and iterative revision. In the thesis, three-dimensional computer modeling (3D CAD) was chosen to produce the 3D computer aided design models of the actual product. Within the thesis, the 3D CAD model was used for four purposes. Firstly learning, the prototype is used as a learning tool. The designed part can be adjusted as many times as it is required until the result is satisfactory. The estimation can be made to test the outcome and generate discussion for further development. Secondly communication, the prototype shows the visual of the product. It not only enhances the communication with the customer and investor, but also the manufacturing process as well. The three-dimensional CAD model is much easier to understand than a sketch or drawing. Thirdly integration, the 3D CAD model can demonstrate the assembly of the product to ensure that the components and sub-systems fit together as designed. Finally as milestones, if the prototype fulfills the requirements, this can be used as a yardstick at various stages of production to ensure that each stage of the development of the product goes well and will perform as expected. (Ulrich & Eppinger, 2008)
Chalmers University of Technology
Chapter 4: Concept development 9 4. Concept development This chapter describes the background of the first prototype, identifies the problem and lists the functions that needed to be improved. The product development process is also outlined in this chapter starting with establishing the required specifications, functional analysis, concept generation of the different functions, evaluation and outcome concept. 4.1 The existing prototype The instrument’s project stemmed from the product development project. The purpose of the project was “To help the company realize their vision of rationalizing the prospecting process and to design a concept for a basis for further development.” (Bragsjö, Halonen, Johansson, Krpo, Sernevi, & Smajic, 2010). The project includes market analysis that was conducted around four main areas: politics, economics and social issues, technology in prospecting, possible market segments and competitors and their solutions. The final concept of the project was called Mobile Automated-Drill Core Analytical Technique (MADCAT) which combined the concepts of different sub-functions, namely, chassis, user interface, analyzing, core-handling, environmental protection and positioning of components. (Bragsjö, Halonen, Johansson, Krpo, Sernevi, & Smajic, 2010) The company decided to produce a single core scanner instead of a core box scanner to reduce complexity of the system. The final concept of the development project differs from the prototype in the following aspects; the number of input samples, sub-system and components because at that time the development team hadn’t gained the knowledge about the components and sub systems yet and there existed a lot of unclear information and uncertainty. Furthermore, the technology was nonexistent or not well known yet. The choices the development team made for the final concept was the best solution for the situation at that time. The existing prototype is a single core scanner. It can only analyze rock samples that are smaller than 0.5 meters at one time. The instrument uses an X-ray fluorescence (XRF) technique for the analyzing process. XRF is a non-destructive analytical technique used to analyze rocks, minerals, sediments and fluids. According to Beckhoff, Kanngießer, Langhoff, Wedell and Wolff, the XRF principle can be explained as follows. When the inner electron is excited by photons in the X-ray, the atoms becomes unstable and an outer electron moves from a higher energy level to fill the gap of the missing inner electron. During the process, energy is emitted due to the energy difference between the two shells. The released spectrum reveals a number of characteristic peaks that can be used to detect the identification of the elements that exist in the sample. During the study phase, the component structure was created to make an understanding of the entire components. The main sub- systems of the prototype are power system, motor system, X-ray system, cooling system, controlling system or PLC, chassis and user interface. The aim of studying the sub systems and its components is to find an answer to these questions. “What are the functions of the components?” and “Why does the instrument need it?” These questions are addressed by studying the manual of each component and then making a components list by categorizing the components into sub systems and listing the important requirements and specifications of each sub system. In addition, a list of pros and cons of each component has been made. Some components might not be suitable for the next
Chalmers University of Technology
10 Chapter 4: Concept development product, such as a high voltage generator with improved technologies, a smaller model with higher power is available and also a laser sensor that can be useful for the developed product was found since producing the first prototype. The cooling system of the prototype is huge and it would probably be better to change it in the developed product of which equipment size is a crucial requirement. The purpose of the instrument is to analyze rock sample which in the prototype case is only a single core at a time, but the next machine is aiming to analyze multiple samples in a standard core box. Figure 4.1 displays the single core scanner. Figure 4.1: The first prototype 4.2 Requirements specification The requirements are based on the requirements of the existing prototype and the requirements of its sub systems and components. The problem is that the requirements of the existing prototype has not been updated and are unclear because of its lack of detail, hard to measure and the limit values are not appropriate. The requirements list of the developed instrument was formulated with additional information and the knowledge gained from the prototype. Furthermore the requirements of the new function of the instrument were added into the requirement list. The big difference between the existing prototype and the new machine is that instead of a single core scanner, the new one can scan multiple samples in a standard core box (See Figure 4.2), which means that the developed equipment is a lot more complex than the existing prototype. It has new functions, such as loading and unloading the core box and the movement of the X-ray components. Another issue is the company wants the new instrument to be compact by putting every component inside the machine and it must be easy to transport. Hence, the new design is suitable to be used at the drill site. For the new product, the new requirements were added and Figure 4.2: The standard core box some of the target levels were changed in order to support the new functions and components of the instrument. After discussions with the company, the first draft of the required specifications was created. The requirements were not complete because some of the values were unknown and some requirements needed to be discussed further in the next phase of the development process. For example, the performance of the
Chalmers University of Technology
Chapter 4: Concept development 11 instrument such as the analytical speed and analysis accuracy depend on the type of material that the user is looking for. The different settings of the searched material can change the speed of the analytical process. Also some requirements are vague and hard to interpret. Another issue that is significant for the instrument is safety with regards to radiation. For this, both Swedish and EU legislations need to be studied so that it can be implemented effectively. The list of requirements consists of the type of requirement, either demand or wish, the requirement statement, target level, validation and evaluation/verification. Table 4.1 shows the examples requirements that have high priorities for the design criteria. The full required specifications can be found in appendix B. Table 4.1: The requirements Requirement Description Samples should not move If the samples move during the analyzing process, it will affect during the analyze process the quality of the result. The user can check the position of the sample from the photo. Sound level The instrument should not have a sound level more than 90dB as this is limited by work place regulation. The sound level can be measured by using a dB meter. Operations manual The machine should have a user’s manual to help the user understand how to set up and execute the processes. Calibration system To help calibrate the instrument in order to get the best accurate result. The system should detect the position of the sample. Internal storage The instrument should have an internal memory used for recoding the result, position and photo of the sample. Easy data transfer with A security system that allows only an authorized person to get authorization lock access to the data. Ergonomic The working process should consistent with the regulation AFS 1998:1 Belastningsergonomi Machine reliability The reliability rate should be within 96% of operating time to fulfill the customer’s expectation. The analyzing process The sample before and after the analysis should be in the same should not affect the state. sample Machine protection During transportation it is important that the machine is suitably protected as the sensitive components could get damaged. Portability It is essential that the machine can be transported to the work site. Rock sample (core box) The machine should be able to input a standard core box and scan the rock samples in both cylinder and half cylinder form, either small or large iron core. Environment Protection from the environment in terms of inside temperature, humidity and water that leak from the core box. Figure 4.3 shows the environment of drill site.
Chalmers University of Technology
14 Chapter 4: Concept development a) Rotation Loading and unloading the sample by rotating the feeding disk. b) Lift up The sample cabinet fits under the machine. The core box is then lifted up to the start position. Unloading happens on the other side of the machine c) Video player Insert the core box horizontally inside the machine d) The oven The feeding hatch opens like an oven door e) Sledge As the hatch swings up, the sledge slides out f) Drawer The hatch slides out like a drawer Figure 4.5 a) - f): Loading and unloading idea solutions. 4.4.2 Movement of the X-ray components and core box The movement of the X-ray components and core box are complicated aspect because the mechanisms of these functions can affect the size of the instrument. The purpose of the thesis states clearly that the developed instrument should be compact and as small as possible. According to the requirements, the size of the machine is limited by the width of a standard door, 80 cm. The height is 75 cm, being the same as the existing prototype. The length of the machine depends on the core box size and the mechanism of this function.
Chalmers University of Technology
Chapter 4: Concept development 15 The movement of the X-ray components and core box should be consistent and work well together. The scanning area should cover the entire sample in the core box. Three possible solutions were realized. In later stages, the two of the solutions were eliminated because it was realized that the core box cannot move along the X-axis. Hence, there was only one option for this function. An animation of the movement of the X-ray equipment and core box was made by using Autodesk inventor for visualization and presented to a big Swedish mining company and potential customers. 4.4.3 Transport solution For the transport solution, the first discussion with stakeholders focused on how to deliver the instrument to the customer safely. Since the machine consists of sensitive components, it is important to make sure that the instrument before and after transport is in the same state, and be able to set up and run as fast as possible. In later discussions, it came out that the possible solution for delivering the instrument from point to point, for example between drilling site or from the company’s office to working location is to use sub-contractors because the instrument is of very high value. Sub-contractors will be responsible for any damages to the machine during transportation. The next topic within transportation is how to protect the instrument from damage during transportation and how to place the instrument inside the vehicle. The transport solution focuses mainly on outer protection and securing the instrument. The special care must be taken to protect the window and the panel and securing the instrument in the vehicle. The solution was inspired by the protection process from different ideas such as the fire truck that uses roll up doors to protect the inside equipment and the container solution that has special attachments whereby the instrument will be secured. Another possible solution would be to make forklift pockets for the forklift to be able to move the instrument. 4.4.4 Maintenance The maintenance list was created by listing the maintenance requirements of each component in order to design the layout for further processes. The component manuals are the significant source of information. The dimensions and technical data, such as operating temperature can be found in these manuals. The maintenance list consists of the list of components in each different sub system, the model of component, weight, operating temperature, storage temperature, humidity, list of maintenance activities and installation notes. The list does not only consider the components of the first prototype but also includes the new possible components that might be involved in the new instrument. The maintenance list gives the information about how often the user needs to access the components. Some sensitive components need special treatment and more dedicated attention than others. Some examples of maintenance would be:  The X-ray tube. In general the glass surface should kept clean and dry at all times, the cooling circuit inside the tube needs to be inspected and cleaned and the X-ray tube itself has a specific life span and needs to be changed.
Chalmers University of Technology
16 Chapter 4: Concept development  The cooling system requires the checking of the water level and heat exchanger. The water needs to be refilled and the filter of the pump are checked and cleaned every 3 months. The maintenance information will affect the components’ layout because the components that require frequent maintenance and regular checking should be placed in the position that’s easy to access by the user. Regarding the non-disclosure agreement, the information within the maintenance list cannot be disclosed because it contains a list of the components, model and supplier and other sensitive information. 4.4.5 Environment The main goal of this thesis is to develop the instrument for field use. To achieve that goal the instrument should withstand the harsh environment of the mining and exploration field. The factors that can affect the performance of the instrument or jeopardize the operation are temperature and humidity inside the machine, because the components of each sub-system can withstand a different range of temperature and humidity. For example, the operating temperature of the high voltage power supply is between -40 to 85 degree Celsius while the X- ray tube can operate at 5 to 40 degree Celsius. The temperature inside the instrument should be adjusted to within a suitable range so that all the systems can operate properly. The solutions of this problem came from different ideas. Some ideas were based on the solutions from other machines such as a heater, fan and dehumidifier to control temperature and humidity or a single component to control both temperature and humidity would be ideal. Other ideas were to use air curtains or controlled air flow to prevent the outside air from contaminating air inside the instrument. 4.4.6 Components layout The idea behind the layout was to fit all components inside the instrument. In the first prototype, the sub systems were located outside the machine including electronic storage cabinet, cooling system and High voltage generator. Only the analysis components were placed inside the machine. In the next product, the functions are more complicated than the prototype which means it has more components. The challenge of the design was to limit the size of the instrument. Other factor that might affect the layout was the sensitivity of the sub systems. For instance, a high voltage generator will create a magnetic field that might affect other electrical components during the operation. Initially two component layouts were proposed to the stakeholders. Autodesk inventor was used to create a visualization of the layout. At this stage, the problems of both layouts were not yet realized. However, throughout the development process the layout has been redesigned and improved many times. The details will be discussed in the next chapter, Refinement. Another issue regarding components’ layout was the design of the X-ray equipment. Measurements of the X-ray parts were conducted in order to get the dimensions of the components. 4.4.7 Safety The safety of the machine is really important. This aspect was considered to be the first priority for the design factor. The safety function of the instrument was based on the Swedish
Chalmers University of Technology
Chapter 4: Concept development 17 Radiation Safety Authority’s rules regarding usage of industrial equipment that contains closed radiation sources and X-ray tubes. The safety functions included a warning light, alarm system and security system. For the warning light, the user should be able to see the light from a 360 degree view around the machine. The possible solutions were to either put one light on top of the machine or install four lights on each corner of the machine. The alarm system involves an emergency stop, status light or alarm sound. Lastly, the security feature was to prevent the user from get exposed to the X-rays. In other words, the door and hatch of the components that connect to the analysis cage should be securely locked as to avoid X-rays getting out. The solutions for this include a key lock, magnetic lock, door sensor and to use a key to start the instrument. 4.4.8 Outer design The design of the front door that is used for maintenance of the components inside the analysis chamber such as the X-ray tube was inspire by the doors of a bus, garage door etc. Examples of the possible solutions are shown in Figure4.6. a) b) c) d) ope Figure 4.6 a) - d): Possible front door solutions The cover of the instrument was designed by Niklas Arthursson, Product designer. The design was based on the idea solutions stemming from the idea generation process. The requirements for the outer design are the instrument should be strong, durable and compact, but still be visible as high-tech equipment. After discussions with the stakeholders in this project, the final design of the outer parts was selected. Figure 4.7 shows the final design of the instrument. Figure 4.7: Final design of the instrument
Chalmers University of Technology
20 Chapter 4: Concept development positioned lights because it passed the safety criterion which is more serious than the cost criteria. The alarm warning function was excluded during this process. Even though the purpose of this function is good, it was not essential for the first development of the instrument. 4.5.3 Kesselring matrix The Kesselring matrices, introduced in chapter 3.4.3, were applied after the Go/No-Go screening. The result from this process will be integrated as the outcome concept. The process of selection begins with the meeting with the company’s representatives to deal with the specifications requirements. The problem is that all the requirements need to be prioritized. By rating each requirement this can help solve the selection process. A rating was given to each requirement depending on the importance of the requirement. Hence: 5 = very important 4 = between medium and very important 3 = medium important 2 = between medium and not important 1 = not important Note that a rating of 1 doesn’t mean that the requirement will not add value to the machine. The solutions that were evaluated by the Kesselring matrices are loading and unloading the core box, environment control and the front hatch. Firstly the method of feeding the core box into the machine, the possible solutions are: video player, the oven, sledge and drawer. When evaluating each possible mechanism, a rating system was applied to each of the following criteria: ergonomics/safety, reliability, performance (speed), size, maintenance, environment (temperature) and cost, the result shows that the video player option has the maximum total weighted value. This means that the best outcome concept for the feeding function is the video player solution. The example of the Kesselring matrix: Loading and unloading the core box is shown in Table 4.4. Secondly the environment control function, the selected criteria focus on controlling the inside temperature, humidity control and reliability of the instrument. Finally the front hatch, the main criteria for each function depends on the selected requirements. Some requirements are significant in all functions because it can affect the status of the entire system such as reliability and maintenance. Table 4.4: Kesselring matrix: Loading and unloading the core box Video player The oven Sledge Drawer Weight Weighted Weighted Weighted Weighted Criteria Value Value Value Value (W) Value Value Value Value (V1) (V2) (V3) (V4) (W*V1) (W*V2) (W*V3) (W*V4) Performance (Speed) 0.11 5 0.53 4 0.42 2 0.21 4 0.42 Weight 0.08 5 0.39 3 0.24 3 0.24 5 0.39 Size 0.11 4 0.42 3 0.32 3 0.32 4 0.42 Ergonomic/Safety 0.13 4 0.53 2 0.26 1 0.13 3 0.39 Reliability 0.13 5 0.66 3 0.39 2 0.26 4 0.53 aesthetic design 0.08 4 0.32 3 0.24 3 0.24 4 0.32 Maintenance 0.11 3 0.32 4 0.42 4 0.42 3 0.32 Control Environment 0.11 4 0.42 2 0.21 2 0.21 4 0.42 (Temp.) Production 0.05 5 0.26 3 0.16 3 0.16 4 0.21 Cost 0.11 4 0.42 4 0.42 3 0.32 4 0.42 Total 1.00 4.26 3.08 2.50 3.84
Chalmers University of Technology
22 Chapter 4: Concept development Figure 4.8: Outcome concept The outer design consists of the front hatch, feeding hatch and the touch screen for human – machine interaction. The front sides of the instrument are protected by the roll up door. The edge and the corners of the machine are mounted with bumpers to absorb any impact that might happen during transportation. With this container concept solution, the instrument can be placed steadily inside the vehicle. For the maintenance, the tray is a good solution to prevent dust and dirt from the samples, getting inside the machine. The airflow control and air filter help collect the dust and purify the air inside the machine. Because the instrument will be located near drill sites, it is possible that the ambient temperature would be extremely low, for example -40 degrees as in the case of the mine in Kiruna in the north of Sweden. In order to ensure that the instrument is operating in this temperature the temperature controller is a good solution. The heater can increase the temperature inside the instrument until it reaches the operating temperature. The dehumidifier can be used to adjust the air humidity. The layout of the machine was designed with the aim to make it as compact as possible, but still fit all the sub-systems including the X-ray components, electronic storage cabinet, cooling system, high voltage generator and computer cabinet. The designed layout also accommodates the movement of the X-ray components and core box. For the safety aspect, which is the most crucial criteria of instruments using X-ray equipment, the EU legislation clearly states that portable equipment that contains a closed radiation should have a well displayed warning light which is lit only when the radiation is present. By putting the warning lights on the corner of the instrument, it helps fulfill the legislation’s requirement. If an unexpected event occurs, the instrument has an emergency stop button to end the process immediately. The most serious issue is the radiation safety. With regards to safety, the main design goal was to prevent the exposure to the radiation. The user is not allowed to handle the X-ray components while the equipment is scanning in any circumstance. The key lock, magnetic lock, door sensor and the use of a key to start the machine not only solves the radiation safety problem, but also limits access to the control panel, sub-system cabinet and sensitive components to an authorized person only. The magnetic lock was installed to secure the front hatch and feeding hatch. The lock responses to the safety PLC
Chalmers University of Technology
24 Chapter 5: Refinement 5. Refinement During the development of the final product, adjustments based on the evaluation methods always need to be made to improve the concept and to combining the good features of some alternative solutions that can be adapted for the product. This chapter will describe the refinement of the requirements specification and the refinement of the final product. 5.1 Refinement of requirements specification The requirements of the instrument were modified during the development process either because of the hidden requirements that hadn’t been realized at the beginning of the thesis, or new knowledge and information that was acquired during the process. These improvements include clarifying the requirements’ statement, selecting a more accurate target level, enhancing the method of validation and evaluation, rating the requirement and adding the new requirements. The specifications were reformulated for the purpose of justifying the requirements. For instance, the requirement that starts with, “Easy to …” such as, “Easy to transport to the mining and exploration site”, “Easy to change between transportation medium” and “Easy to load/unload the machine into the vehicle” were difficult to define whether it is easy or not. After the discussion the requirements were changed to “Possible to…”. The changed requirements fulfill the basic understanding that the instrument should be able to do as stated in the requirements. The target level was a tricky aspect because if the target level was too high to achieve, it might create a problem with the actual production of the machine. The preliminary solution was to set the target within a required range and the demand and wish target level. After carefully consideration the target level was changed to a more appropriate and accurate value for example, the analytical speed is 1mm/s to 10mm/s depending on the type of material, the analysis accuracy would be 100% accurate for the demand. The validation and evaluation methods were explored to find the ways to improve the methods and procedures. For example to validate the mechanical robustness the instrument should be compared to other machines and equipment that are used in the mining industries. For the safety function, it is stated in the initial requirement that the instrument should automatically shut down the X-ray tube when the hatch is open, but when considering the maintenance aspect, it might occur that the user needs to open some hatch, such as the computer cabinet or the electronic storage cabinet, while the machine is operating. The final product considers both the safety and maintenance aspects by designing an inner cage that prevents radiation leakage. In order to comply with this design, the requirement was added that the safety sensor be applied only to the hatch that leads to the X-ray tube so the user should still be able to open the electronic storage and computer hatch while operating. The main specification change that affected the design is the size of the core box. The original intention of the design was to have a core box size of 1042 mm x 378 mm x 50 mm, but during the development, new information came to light that the standard size of the core box varies depending on the country that uses it. Thus the new dimension of the core box is, 1100 mm x 362 mm x 50 mm making the length of the new core box longer than the first one.
Chalmers University of Technology
Chapter 5: Refinement 25 5.2 Refinement of the concept The result from the evaluation methods is only the concept. To make it more feasible the refinement was needed to fill in the missing detail. Following are the concerns of the instrument and the solutions to overcome the problems. Some of the concerns were discovered during the 3D CAD modeling process. 5.2.1 Loading and unloading the core box and feeding hatch For the method of loading and unloading the core box, the mechanism called “video player” was chosen. The concerns of this system are the lack of detail about the moving parts, how to open the hatch and how it works. The modified version of the “video player” integrates the strength of the “Drawer” concept. The improved concept satisfies all the concerns. The main features of the mechanism are the movement of the feeding tray and the hatch. The hatch was designed as a lifting door which could open upwards or downwards. After investigating both options the conclusion was that the best solution is to open downward because the top of the hatch could be used to attach the roller or other material to support the tray when loading the core box. The hatch should have a physical lock as well as the mechanism to prevent it from falling down. Research was conducted to find an example of a product that used such a system. The suppliers that can provide a similar type of product were found. The problem of this solution was realized later. Because of the size of the feeding hatch, it is not only impossible to make a lifting hatch that has a counter weight mechanism, but also extremely expensive for a motorized solution. After discussions with the stakeholders, the decision was made to reduce the complexity of the design by changing the lifting hatch to a normal swing hatch that opens downward. The decision not only solves the feeding hatch problem but also simplifies the space problem which is very tight inside the machine. 5.2.2 Cooling system In the final product, the cooling system is placed in the back of the instrument next to the electronic storage cabinet. The drawing from the supplier indicates that if the cooling system is positioned in the back, the air inlet and outlet will be placed on the backside of the equipment as well. A ventilation problem might occur because the machine will be located close to the wall where there is restricted air flow. An alternative solution was to place the cooling system on the side of the machine. The only concern with this was the gap above the cooling system will be really tight because during the analysis the X-ray components are moving to the left ends and right ends of the instrument. The problems were solved when a new model of heat exchanger was discovered, which is very thin but longer. The decision was to put it on the top of the machine. Even though the total height of the machine will be increased, but when considering the advantage of having the air inlet and outlet on the top of the machine, this solution becomes the most suitable choice. The space at the back of the equipment can be used to install the water tank and pump because it is easier for maintenance when the water tank is separate from heat exchanger, for instance to refill the water, drain the water out of the system and cleaning. With this solution, the electrical cabinet can be expanded because the cooling system cabinet will use less space than in the first layout. 5.2.3 Core box dimension The most critical issue was the discovery of the new dimensions of the core box. Since the machine was designed to use the standard core box (1042 mm x 378 mm x 50 mm), the new box is 1100 mm in length which can jeopardize the design of the machine. The reason behind
Chalmers University of Technology
Chapter 6: Result 27 6. Result In this chapter, the final design of the instrument and the limitations of the design will be described. The topics focus on the feasible and ready to produce concept, the three- dimensional computer modeling (3D CAD) and the problems that arose during the 3D CAD modeling. 6.1 Final Design After the refinement process, it is time to finalize the concept and design the components. The main parts of the instrument that were designed in this thesis consist of the profile structure, inner cage, cabinet, analysis equipment, feeding components, front hatch, feeding hatch and outer plate. The 3D CAD modeling was chosen as a prototyping tool to create the 3D model and production drawing. 6.1.1 Profile structure The profile structure is the aluminum frame that supports the inner cage and the outer plate. The total height of the instrument was expanded because the legs. The reason why the instrument needs the legs is for the transportation. The machine requires a gap between the body and the ground at least 130 mm in order to fit the forklift forks. Because the main function of the frame is to receive the load from the equipment, the beam that is used as the main structure is thicker than the rest. Triangular profiles were placed at the corners of the structure for aesthetics and for the installation of the warning lights on the top of the profile. Figure 6.1 shows a rendering of profile structure created by using the computer-aided design software, Autodesk Inventor. Figure 6.1: The profile structure 6.1.2 Inner cage The inner cage was designed to separate the X-ray beam from the user and other components. It is important that the primary beam of the X-ray cannot be reached by anyone at any time. The crucial factor that needs to be considered when designing this part was the radiation safety. The thickness of the metal sheet required is realized when the outer plate is included, the total thickness is sufficient to shield against radiation from the X-ray tube. The problem is when the user opens one of the cabinets during operation the total thickness of the metal plate is then reduced. The design solution is to use the thicker plate on the cabinet side. In