Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Experimental Detection of Organised Motion in Complex Flows with Modified Spectral Proper Orthogonal Decomposition
Next Article in Special Issue
Physics-Informed Super-Resolution of Turbulent Channel Flows via Three-Dimensional Generative Adversarial Networks
Previous Article in Journal
Compressible and Viscous Effects in Transonic Planar Flow around a Circular Cylinder—A Numerical Analysis Based on a Commercially Available CFD Tool
Previous Article in Special Issue
Turbulence Modeling for Physics-Informed Neural Networks: Comparison of Different RANS Models for the Backward-Facing Step Flow
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DropletAI: Deep Learning-Based Classification of Fluids with Different Ohnesorge Numbers during Non-Contact Dispensing

1
Laboratory for MEMS Applications, Department of Microsystems Engineering—IMTEK, University of Freiburg, Georges-Koehler-Allee 103, 79110 Freiburg, Germany
2
Hahn-Schickard, Georges-Koehler-Allee 103, 79110 Freiburg, Germany
3
Department of Computer Science, University of Freiburg, Georges-Köhler-Allee 106, 79110 Freiburg, Germany
*
Author to whom correspondence should be addressed.
Fluids 2023, 8(6), 183; https://doi.org/10.3390/fluids8060183
Submission received: 17 May 2023 / Revised: 8 June 2023 / Accepted: 9 June 2023 / Published: 17 June 2023
(This article belongs to the Special Issue Machine Learning and Artificial Intelligence in Fluid Mechanics)

Abstract

:
The reliable non-contact dispensing of droplets in the pico- to microliter range is a challenging task. The dispensed drop volume depends on various factors such as the rheological properties of the liquids, the actuation parameters, the geometry of the dispenser, and the ambient conditions. Conventionally, the rheological properties are characterized via a rheometer, but this adds a large liquid overhead. Fluids with different Ohnesorge number values produce different spatiotemporal motion patterns during dispensing. Once the Ohnesorge number is known, the ratio of viscosity and surface tension of the liquid can be known. However, there exists no mathematical formulation to extract the Ohnesorge number values from these motion patterns. Convolutional neural networks (CNNs) are great tools for extracting information from spatial and spatiotemporal data. The current study compares seven different CNN architectures to classify five liquids with different Ohnesorge numbers. Next, this work compares the results of various data cleaning conditions, sampling strategies, and the amount of data used for training. The best-performing model was based on the ECOmini-18 architecture. It reached a test accuracy of 94.2% after training on two acquisition batches (a total of 12,000 data points).

1. Introduction

Droplet dispensing is the process of ejecting discrete fluidic volumes with two immiscible fluids [1]. The ejection process can happen in various droplet breakup regimes, one of them being an on-demand breakup. The dispense is performed by providing the fluid with enough energy as a short actuation pulse in the form of thermal fluctuations or mechanical vibrations to overcome energies holding the fluid back, such as capillary forces.
When this actuation pulse is applied, the droplet takes a teat shape due to the positive pressure and the opposing surface tension where the fluid touches the orifice [2]. Next, a tail is formed, caused by the inertia of the liquid, and this tail is broken by the upward pull of the negative pressure created. After leaving the nozzle, the droplet continuously changes its shape from the initial cylindrical-like shape to wobbling between tear-like and spherical shapes, until it becomes stably spherical due to the surface tension of the fluid. Hence, during the initial phase of the ejection, the droplet undergoes a shape evolution before reaching a stable spherical state, i.e., it follows a specific spatiotemporal trajectory. For non-contact dispensers (such as PipeJet P9, BioFluidix GmbH, Freiburg im Breisgau, Germany [3]), this spatiotemporal motion pattern depends on various factors: the geometries involved (such as the characteristic length (D) of the dispenser), the actuation dynamics (such as the stroke length (SL) and downstroke velocity (DV)), the rheological (such as viscosity) and physical properties (such as density) of the liquid being dispensed, the surface effects, and the environmental variables (such as temperature and relative humidity). This primary droplet can also be followed by secondary droplets known as satellites.
To simplify the modeling of the ejection events, characteristic numbers can be used. These numbers are dimensionless quantities and hence systems with the same characteristic numbers behave similarly regardless of factors such as their scale. The dimensionless numbers which are interesting for the droplet ejection process are the Weber number (We) and the Ohnesorge number (Oh) [1]. The droplets will follow different spatiotemporal trajectories when ejected either with a different We and/or with a different Oh. Figure 1 shows the output of a simulation conducted wherein fluid is ejected at various Oh [4].
We is the ratio between the kinetic energy of a fluid and its surface energy [1]. Therefore, it depends on the actuation parameters and the liquid properties, whereas Oh relates the following: internal viscous dissipation to inertia and surface tension. Therefore, it depends on the liquid properties and the characteristic length. The relation between We, Oh, and the Reynolds number (Re) is given in Equation (1)
O h = W e / R e = η / ρ · D · σ
where η is the dynamic viscosity of the liquid, ρ is the density, D is the characteristic length, and σ is the surface tension. The spatiotemporal trajectories formed by the ejected droplets can be imaged using a high-speed camera.
Even though mathematical models exist to create these droplet trajectories from the initial conditions [5], solving the inverse problem (i.e., estimating the initial conditions from the trajectories) is not trivial. Nonetheless, such a solver is useful in practical scenarios. It could be used, for instance, to adjust dispenser parameters on the fly as the liquid properties change due to temperature fluctuations. As neural networks are universal approximators [6], these can be applied to solve the aforementioned problem. Convolutional neural networks (CNNs) are great tools for capturing the spatial and spatiotemporal content [7]. Hence, these can be used to identify liquids with different Oh from the captured images and image sequences (videos) of the dispensed droplets.
Convolutional neural networks (CNNs), which belong to the class of machine learning algorithms known as deep learning (DL) [8], build complex representations by taking simpler representations as input in an end-to-end manner. This ability to automatically build new constructs on top of existing representations gives DL the ability to extract the non-obvious complex representations from the underlying data. This can then be used to perform tasks such as distinguishing between different classes present in the data, i.e., classification. The training for this classification process can be conducted in a supervised fashion if the type of liquid in the image, i.e., the label, is known.
As droplets with different Oh follow different spatiotemporal trajectories, they may form unique shapes. This indicates that neural networks with only 2D extractors may be enough and 3D information processing may not be needed. To test this, different CNN architectures with pure 2D convolutions, pure 3D convolutions, or 3D–2D mixed convolutions were compared.

2. Data Acquisition

Data acquisition is one of the major tasks in the DL pipeline. To fulfill the need of acquiring a large amount of labeled data, an automated data acquisition setup, named TestRig, was used. For each droplet dispense, a sequence of 250 frames at 6806 frames per second was recorded. Figure A2 shows an example sequence. During the data acquisition process, various sources for shortcut learning [9] such as poor lighting, poor focus, tilting of the dispenser nozzle, etc., were avoided. The following topics are discussed in the following subchapters: hardware and software of the TestRig (Section 2.1), the test liquids used (Section 2.2), as well as hardware configuration and acquisition protocols (Section 2.3).

2.1. Hardware and Software of the TestRig

Figure 2 shows the TestRig. It was built on top of an optical breadboard. The droplets were dispensed using a nanoliter non-contact droplet dispenser (BioFluidix PipeJet®, Freiburg im Breisgau, Germany [3]) mounted on a three-axis precision stage. The nozzles had an inner diameter of 500 μ m . The falling droplets were imaged as shadows by positioning the dispenser nozzle between a light-emitting diode and a high-speed camera (MotionBLITZ EoSens® mini2, Mikrotron GmbH, Unterschleissheim, Germany). To reduce the influence of convection on the trajectory of the droplets, the setup was covered with walls and a lid.
The liquid pressure was regulated during the acquisition using a pressure control module (ActivePCR, BioFluidix GmbH, Freiburg im Breisgau, Germany). Additionally, the weight of the falling droplets was measured using a micro-weighing balance (SC2 Ultra-Microbalance, Sartorius AG, Goettingen, Germany) placed under the orifice, and the ambient temperature, pressure, and humidity were logged with a thermo-hygrometer.
The TestRig was controlled via an in-house developed software called ♯Drop [10]. It was used to trigger the camera, configure the parameters of the PipeJet® actuation, and log mass, ambient temperature, relative humidity, and pressure readings. Next, MotionBLITZ® Director2 was used to configure and operate the camera. Furthermore, Biofluidix Pipejet® Control software was used to configure and operate the pressure control unit.

2.2. Used Liquids

This study uses five of the eight model liquids developed in [11], covering the whole range of viscosity and surface tension of 646 of the most common in vitro diagnostics reagents. Figure 3 presents the surface tension vs. viscosity for the model liquids. Table 1 summarizes fluid properties, the Oh for a 500 μ m nozzle, and the components shortlisted from [11]. There were little fluctuations in the ambient temperatures (23.2 °C to 28.6 °C) during acquisition, but the resulting error in the Oh calculation (for water) was not significant (0.54 × 10−3) as all the other liquids used in this study have significantly larger Oh.

2.3. Hardware Parameters and Acquisition Protocol

A set of 64 configurations was chosen for driving the Pipejet®. Each of these configurations was repeated 25 times consecutively, leading to a total of 1600 dispenses. Each configuration was a combination of one of the eight SLs (10, 12, 14, 16, 18, 21, 23, or 25 μ m ) and one of the eight DVs (60, 70, 80, 90, 100, 110, 120, or 130 μ m / m s ) with all the DVs sequenced in ascending order with a single SL before moving to a higher SL. An acquisition batch is a set containing one acquisition cycle from all five test liquids. The data acquisition was conducted in three acquisition batches.
Depending on the viscosity of the individual liquids, some actuation configurations did not provide enough energy to eject the liquid. These invalid configurations were filtered out of the dataset. This resulted in 31 to 64 valid configurations based on the liquid, nozzle condition, and environmental conditions.
The camera was configured by defining a frame size with a height of 750 pixels and a width of 144 pixels. The height was experimentally determined to ensure that, even for the drop with the longest tail (produced with the liquid with the highest Oh, liquid E), there should exist at least one frame that contains the entire droplet, from head to tail. Next, the width was obtained by observing the number of pixels covered by the nozzle and multiplying it by 3 to account for the tilting of the nozzle. The configurations resulted in a maximum frame rate of 6806 fps for the current hardware. For each dispense, 250 images were acquired at this frame rate.
The tip of the nozzle was kept in the frame. This was primarily because the droplets have a high degree of spatiotemporal changes when leaving the nozzle, which can be interesting for the classification task. Moreover, the nozzle of known diameter can also be used as a reference object for camera zoom and focus. A shadowgraph of the field of view in the imaging plane is given in Figure 4.
The time difference between two consecutive dispenses was 31 s for the first two acquisition batches to save all the acquired images. It was increased to 33 s in the third acquisition batch to give the dataset a bit of diversity and improve the generalization of the trained model on new, unseen datasets. Gravimetric weight analysis was performed according to the GRM method [12,13] with settling and measuring times of eight seconds each in the first two acquisition batches and which were increased to 15 s in the third to measure the slowest falling droplets accurately.

2.4. First Appearance of Droplet

Each image sequence had 250 images, in which some of the initial frames were empty. The number of empty frames was dependent on the PipeJet® configuration used and was independent of the liquids used in this study. The first frame where the droplet appears was identified by running a data pre-processing script. The first appearance ranged from frame 71 (SL = 25 μ m , DV = 60 μ m /ms)–186 (SL = 10 μ m , DV = 130 μ m /ms) (Appendix A).

3. Deep Learning

3.1. Architectures

In the current study, seven different neural network architectures were compared. These neural networks can be classified based on the type of data stream (pure 2D, pure 3D, or both 2D and 3D) or based on the network depth (shallower or deeper).
There were two pure 2D architectures: ResNet-18 [14] (which contains 18 2D-convolution layers) and ResNet-C (a 14-layer variant of ResNet-18). Both of these were relatively shallow networks. The ResNet (or Residual Network) architecture introduced skip connections that are effective in reducing the vanishing gradients problem of deep networks. There was one purely 3D architecture (3D-ResNet18 [15]). This is a variant of Resnet-18 where the pure 2D convolutions are replaced with pure 3D convolutions to capture the spatiotemporal content of the data. The remaining four architectures contained both the 2D and 3D convolutions. These are ECO Lite, ECO Full [7], ECOmini-18, and ECOmini-C. ECO Lite and ECO Full are the deeper networks used in this study. These were originally made for action recognition on videos, where the relevant information is also spread out over multiple frames. These networks were created on an assumption that, even though the information relevant for classification spans multiple frames, single frames also contain significant information regarding the task.
In both these architectures, the image frames are initially fed to 2D convolution layers. This is achieved using the BNInception architecture until the inception-3c layer [16]. In the ECO Lite architecture, the representations from these individual frames are stacked together and fed to the 3D-ResNet18 architecture. The ECO Full architecture further augments the ECO lite by adding a 2D convolution stream parallel to the 3D-ResNet18 architecture. This 2D stream is made using the BNInception architecture from the inception-4a layer to the last pooling layer. The two steams are fused later in the network by appending the outputs in a single fully connected layer.
The last two architectures, ECOmini-18 and ECOmini-C, were made on a blueprint similar to the ECO Full architecture. In ECOmini-18, the layers up to inception-3c from ECO Full were replaced by the first nine of ResNet-18, and the rest of the layers were used in parallel to 3D-ResNet18, whereas in ECOmini-C, the first seven layers of ResNet-18 replaced the layers up to inception-3c and used the rest parallel to 3D-ResNet18. As these architectures use a ResNet-18 backbone, they were relatively shallower. Table 2 summarizes the seven architectures and the type of convolutions utilized.

3.2. Hyperparameters and Hyperparameter Optimization

While training the neural network models, different hyperparameters such as learning rate (LR), drop out, batch size, etc., were tuned. The choice of optimal hyperparameters was made using a sequential model-based algorithm configuration (SMAC) [17] run for a maximum of 20 epochs. For each training run, SMAC was run for a maximum of 24 h either on an NVIDIA® Tesla® V100 or on an NVIDIA® GeForce® RTX 2080 Ti. The set’s accuracy was obtained from the epoch with the highest validation accuracy. The hyperparameters (and their SMAC limits) were as follows: LR (10−6–10−1), drop out (0.01–0.99), and batch size (8–32).
In addition to the above, a decaying LR strategy was used within each SMAC instance. In this strategy, the LR was reduced to one-tenth of its value if the accuracy did not improve for the “n” number of epochs.The choice of this “n” count was also a hyperparameter (3–10). Next, the networks were provided with an option to retain self-normalization after the dropout layer. This was achieved using a flag that switches between dropout and alpha dropout as a hyperparameter. Lastly, a hyperparameter was used to choose the data augmentation level. Control over the amount of augmentation is required to ensure that the augmented data still represent the actual process of droplet ejection.

3.3. Data Augmentation

Three techniques were used to perform data augmentation: horizontal flipping with a flipping probability of 50%, contrast and brightness jitter with a random value up to 5%, and random cropping with different augmentation levels. There were six augmentation levels (0–5) that represent a maximum reduction in side length due to random cropping and were used as a hyperparameter which was optimized via SMAC. An augmentation level of 0 represented no cropping, whereas an augmentation level of 5 represented a maximum cropping of 10% of the side length, preserving the aspect ratio. The crops were upscaled to the original image size before feeding to the neural network.

3.4. Experimental Setup

The current work was a five-class classification task, trained with stochastic gradient descent, using cross-entropy as a loss function. Acquisition Batch 1 was used for training and Batch 2 for testing. To analyze the effect of adding more data, acquisition batch 3 was also included in the training data. To ensure reproducibility, the GPU and CPU pseudorandom generators were seeded with the same value of 0 for deterministic runs.
The dataset was used either uncleaned (referred to as CV ) or cleaned with CV values of 5% or 2% calculated from the droplet mass values obtained during data acquisition. As a pre-processing step, the grayscale images were shrunk by a factor of 6, resulting in an image size of 125 × 24 pixels. The last pooling layer was made with rectangular kernels with a 4:1 height-to-width ratio.
The input images were obtained by sampling down the acquired videos to n-images (4, 8, 12, or 16) with either the random sampling (RS) [7] or sequential sampling (SS) technique. In random sampling, the image sequence is temporally separated into n-segments of equal size, and one image is sampled from each segment. This strategy helps the network to see the images from the entire temporal domain. In sequential sampling, n-images are sampled sequentially from a predefined starting point. By this strategy, the network is fed with sequential localized information from the region of interest. The starting point for this strategy was selected at a point after the first appearance of the droplet. Therefore, the images can be sampled from the initial region of dispense where there are a significant number of changes in the shape of droplets or these can be from a later part in the sequence where the droplet is mostly spherical. Figure 5 illustrates both sequential and random sampling techniques.

4. Results and Discussions

The first tests were performed on deeper architectures. These networks were trained for 70 epochs. 3D-Resnet18, ECO Lite, and ECO Full had 87.40%, 87.62%, and 89.68% respective accuracies. Extensive testing was not performed on these architectures because they were overfitting. The rest of this section compares the shallower architectures, compares the sampling strategies, analyzes the effect of data cleaning, and analyzes the effect of adding the third acquisition batch. All these tests use eight segments.

4.1. Comparing Shallower Architectures

As there were various reasons (sampling strategy, data cleaning, and the number of batches used for training) which may lead to different accuracies for the same neural network architecture, this section compares the performance of the networks over these variables.
Figure 6 compares the pure 2D architectures on various filtering intensities (expressed as CV thresholds) and sampling conditions. ResNet-C has a lower performance as compared to ResNet-18. This shows that, for the current task, the higher number of layers in ResNet-18 makes it better at extracting the underlying features for every sampling and cleaning condition.
Next, Figure 7 compares ECOmini-18 and ECOmini-C on various CV and sampling conditions. ECOmini-18 and ECOmini-C showed similar performance on the given dataset. This shows that adding the 3D layers with the ResNet-C helps it to extract more useful information from the current dataset.
To analyze the effect of adding a 3D component with a good 2D extractor, a comparison between ECOmini-18 and ResNet-18 is shown in Figure 8. ECOmini-18 and ResNet-18 showed similar performance. This shows that the spatiotemporal trajectories produced during dispensing at the current Oh produce distinct enough underlying frames that the 2D components extracted by ResNet-18 alone are enough for the classification task. A high difference between the Oh leads to very different underlying images [4]; hence, the current 2D extractors can differentiate between the Oh based on the single frames. Hence, for datasets similar to the one used in this study where the difference in the Oh is large, a 2D architecture can be used to make inferences with a single frame.

4.2. Comparing Sampling Strategies

Depending on the non-contact dispensing parameters, the droplets had a different first frame of appearance and were dispensed at different velocities. These lead to a different number of empty frames before the dispense and after the droplet had left the region of interest. Therefore, a random sampling strategy can have a significant number of empty frames. Figure 9 shows a performance comparison between ResNet-18 and ECOmini-18 while using the two sampling strategies.
In the case of ResNet-18, both sampling strategies had very similar performances. As it is a 2D architecture, ResNet-18 does not extract any patterns that exist between the frames. Hence, it is not affected by the type of sampling, whereas in the case of ECOmini-18, sequential sampling outperforms random sampling for the current dataset. This can be because, for the current dataset, the useful information on shape evolution lies in the initial part of the dispenses. To verify this, sequential sampling with different offsets was performed. Figure 10 shows a sequential sampling of eight frames with no offset, a five-frame offset, and a ten-frame offset for ECOmini-18 and ResNet-18.
The neural networks perform the best when fed with images from the initial part of the sequence. Hence, the relevant information is relatively more localized in the initial part of the sequences.

4.3. Effect of Data Cleaning

The data were cleaned based on the CV of the droplet mass obtained within the 25 repeats of the same dispenser configuration. The lower the CV threshold, the lower the amount of remaining data. It is worth noting that CV represents the uncleaned data. Figure 11 compares the performance of the ECOmini-18 and ResNet-18 over different CV thresholds and sampling strategies.
The performance of the neural networks decreases with the decrease in the amount of data as the amount of filtering is increased.

4.4. Effect of Adding New Acquisition Batch

Training on more than one acquisition batch helps the network reduce the shortcut learning that can arise by training on a single batch. The increase in the amount of diverse data helps the neural network to better generalize. Figure 12 shows a comparison between the networks trained on a single acquisition batch and the networks trained on two acquisition batches. The data here are cleaned with a CV threshold of 5%.
The performance of the neural networks increases when the training is performed on acquisition batch 1 and 3 jointly. The ECOmini-18 model obtained by training on acquisition batch 1 and 3 jointly was the best model in the entire study.

5. Conclusions

The current work compares seven different neural network architectures over different sampling techniques, data cleaning conditions, and number of acquisition batches. In conclusion, deeper architectures (such as ECO Full) tend to overfit while finding the Ohnesorge number from dispensed droplet images. A shallower architecture with both 2D and 3D convolutions (ECOmini-18) performs slightly better than a shallower architecture with only 2D convolutions (ResNet-18). Though it has a slightly lower accuracy, inference can be made using ResNet-18 with as little as one image. The results also show that the CNNs with both 2D and 3D convolutions perform better when the frames are sampled sequentially, starting from the frame of the first appearance of the droplet. Next, it can be seen that data cleaning causes a decrease in classification accuracy due to a reduction in the amount of training data. Lastly, it can be seen that training on more than one acquisition batch improves the accuracy of the trained model. ECOmini-18 showed the maximum accuracy of 94.2%, trained on data from acquisition batches 1 and 3 jointly. This study is a proof-of-concept for neural networks to predict Ohnesorge numbers from droplet images. Moving forward, the technique can be further developed into a regression task, where the Ohnesorge number is a continuous scalar, instead of restricted to discrete classes. The CNNs can further be applied to other applications such as predicting the pinch-off time. Furthermore, the current CNNs can be trained in a physics-informed setup, reducing the size of datasets and biases from the experimental data.

Author Contributions

Conceptualization, R.Z., P.K., S.K., P.S., M.Z., and T.B.; Methodology, R.Z., P.K., S.K., P.S., M.Z., and T.B.; software, R.Z., P.K., S.K., P.S., M.Z., and T.B.; validation, R.Z., P.K., and T.B.; formal analysis, R.Z., P.K., S.K., P.S., M.Z., and T.B.; investigation, P.S.; resources, R.Z., P.K., S.K., M.Z., and T.B.; data curation, P.S.; writing—original draft preparation, P.S.; writing—review and editing, P.S., S.K., M.Z., G.M., P.K., R.Z., and T.B.; visualization, P.S.; supervision, S.K., M.Z., P.K., R.Z., and T.B.; project administration, S.K., M.Z., P.K., R.Z., and T.B.; funding acquisition, S.K., P.K., and R.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The code for the work can be found at https://github.itap.purdue.edu/psardana/DropletAI. A sample dataset containing the first 200 dispenses from every acquisition batch for all five fluid classes can be found at https://osf.io/yxqjn/. The authors can provide the full dataset upon request.

Acknowledgments

The authors would like to thank Baden-Württemberg’s HPC for providing resources (BwForCluster NEMO) for a big part of the work. We acknowledge support by the Open Access Publication Fund of the University of Freiburg.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations and variables are used in this manuscript:
CNNsConvolutional Neural Networks
DCharacteristic Length
SLStroke Length
DVDownstroke Velocity
η Dynamic Viscosity
ρ Density
σ Surface Tension
WeWeber Number
OhOhnesorge Number
ReReynolds Number
DLDeep Learning
LRLearning Rate
SMACSequential Model-Based Algorithm Configuration
RSRandom Sampling
SSSequential Sampling

Appendix A. Detailed Dataset Analysis

The complete dataset had 18,550 videos distributed over three acquisition batches and five classes. Figure A1 gives an overview of data distribution over different acquisition batches and classes.
Figure A1. An overview of data distribution over different acquisition batches and classes.
Figure A1. An overview of data distribution over different acquisition batches and classes.
Fluids 08 00183 g0a1
Each video was a sequence of 250 frames with some of the initial frames empty because the fluid takes some time to respond to the actuation pulse applied. Figure A2 shows the entire sequence of 250 frames obtained for Fluid A when dispensed at SL = 21 μ m and DV = 90 μ m /ms.
To perform a more detailed analysis, the distribution of the dispenser configurations used in each class was also observed. A heatmap for the coefficient of variation (CV) distribution is shown in Figure A3.
In the figure, an upper CV threshold (of 50%) was used to make the visualizations clear. The current gravimetric weight analysis is sensitive to the used measurement time. Hence, it produced a high CV for droplets falling with low velocity or dispensing with satellites. The “-” in the image are the cases for which no droplet was ejected when the dispenser was driven at those configurations. Next, a similar analysis for the frame of the first appearance of the droplet was made. The results show that the frame of the first appearance of the droplet remained almost the same (the maximum difference was 1 frame and there was no clear pattern for its origin) irrespective of the liquid used. Figure A4 summarizes the relationship between the dispenser configuration used and the frame with the first appearance of the droplet. This analysis was only conducted for a set of dispenses with a CV of 5% or less.
Figure A2. The sequence of 250 frames obtained for Fluid A when dispensed at SL = 21 μ m and DV = 90 μ m /ms.
Figure A2. The sequence of 250 frames obtained for Fluid A when dispensed at SL = 21 μ m and DV = 90 μ m /ms.
Fluids 08 00183 g0a2
Figure A3. A heatmap for the coefficient of variation (CV) distribution. The heatmap was trimmed at an upper threshold for better visualization.
Figure A3. A heatmap for the coefficient of variation (CV) distribution. The heatmap was trimmed at an upper threshold for better visualization.
Fluids 08 00183 g0a3
Figure A4. The relationship between the dispenser configuration used and the frame with the first appearance of the droplet.
Figure A4. The relationship between the dispenser configuration used and the frame with the first appearance of the droplet.
Fluids 08 00183 g0a4

References

  1. Lindemann, T.; Zengerle, R. Droplet Dispensing. Encycl. Microfluid. Nanofluidics 2013, 1–14. [Google Scholar] [CrossRef]
  2. Wu, H.C.; Lin, H.J. Effects of Actuating Pressure Waveforms on the Droplet Behavior in a Piezoelectric Inkjet. Mater. Trans. 2010, 51, 2269–2276. [Google Scholar] [CrossRef] [Green Version]
  3. BioFluidix GmbH. PipeJet Nanodispenser Kit; BioFluidix GmbH: Freiburg, Germany, 2020. [Google Scholar]
  4. Lindeman, T. Droplet Generation-from the Nanoliter to the Femtoliter Range. Ph.D. Thesis, Albert-Ludwigs-Universität Freiburg, Freiburg im Breisgau, Germany, 2006. [Google Scholar]
  5. Fuster, D.; Agbaglah, G.; Josserand, C.; Popinet, S.; Zaleski, S. Numerical simulation of droplets, bubbles and waves: State of the art. Fluid Dyn. Res. 2009, 41, 065001. [Google Scholar] [CrossRef]
  6. Hornik, K.; Stinchcombe, M.; White, H. Multilayer feedforward networks are universal approximators. Neural Netw. 1989, 2, 359–366. [Google Scholar] [CrossRef]
  7. Zolfaghari, M.; Singh, K.; Brox, T. ECO: Efficient Convolutional Network for Online Video Understanding; Springer: Berlin/Heidelberg, Germany, 2018; Volume 11206, pp. 713–730. [Google Scholar] [CrossRef] [Green Version]
  8. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  9. Geirhos, R.; Jacobsen, J.H.; Michaelis, C.; Zemel, R.; Brendel, W.; Bethge, M.; Wichmann, F.A. Shortcut learning in deep neural networks. Nat. Mach. Intell. 2020, 2, 665–673. [Google Scholar] [CrossRef]
  10. Liang, D.; Zhang, J.; Muniyogeshbabu, T.G.; Bammesberger, S.; Tanguy, L.; Ernst, A.; Koltay, P.; Zengerle, R. Multi-principle droplet calibration technology. microTEC Suedwest 2012. Available online: https://www.imtek.de/data/lehrstuehle/app/dokumente/conferences-pdf/conferences-2012/liang-multi-principle-droplet-calibration (accessed on 8 June 2023).
  11. Losleben, N. Contact-Free Dispensing for In-Vitro Diagnostics: Challenges of the Reagent Diversity on the Performance of Appropriate Dispensing Technologies. Ph.D. Thesis, Albert-Ludwigs-Universität Freiburg, Freiburg im Breisgau, Germany, 2015. [Google Scholar]
  12. Liang, D.; Steinert, C.; Bammesberger, S.; Tanguy, L.; Ernst, A.; Zengerle, R.; Koltay, P. Novel gravimetric measurement technique for quantitative volume calibration in the sub-microliter range. Meas. Sci. Technol. 2013, 24, 025301. [Google Scholar] [CrossRef] [Green Version]
  13. ISO 23783-2:2022; Automated Liquid Handling Systems—Part 2: Measurement Procedures for the Determination of Volumetric Performance. International Organization for Standardization: Geneva, Switzerland, 2022.
  14. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  15. Hara, K.; Kataoka, H.; Satoh, Y. Can Spatiotemporal 3D CNNs Retrace the History of 2D CNNs and ImageNet? In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 6546–6555. [Google Scholar]
  16. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  17. Hutter, F.; Hoos, H.H.; Leyton-Brown, K. Sequential Model-Based Optimization for General Algorithm Configuration; Springer: Berlin/Heidelberg, Germany, 2011; Volume 6683, pp. 507–523. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Droplet ejection simulation for various Oh: (a) 0.066, (b) 0.656, (c)1.312, (d) 1.967 [4]. The figure is reproduced with permission from the author.
Figure 1. Droplet ejection simulation for various Oh: (a) 0.066, (b) 0.656, (c)1.312, (d) 1.967 [4]. The figure is reproduced with permission from the author.
Fluids 08 00183 g001
Figure 2. TestRig. It consists of a dispenser (BioFluidix PipeJet®, Freiburg im Breisgau, Germany), a high-speed camera (MotionBLITZ EoSens® mini2, Mikrotron GmbH, Unterschleissheim, Germany), a pressure control module (ActivePCR, BioFluidix GmbH, Freiburg im Breisgau, Germany), a balance (SC2 Ultra-Microbalance, Sartorius AG, Goettingen, Germany), and a thermo-hygrometer.
Figure 2. TestRig. It consists of a dispenser (BioFluidix PipeJet®, Freiburg im Breisgau, Germany), a high-speed camera (MotionBLITZ EoSens® mini2, Mikrotron GmbH, Unterschleissheim, Germany), a pressure control module (ActivePCR, BioFluidix GmbH, Freiburg im Breisgau, Germany), a balance (SC2 Ultra-Microbalance, Sartorius AG, Goettingen, Germany), and a thermo-hygrometer.
Fluids 08 00183 g002
Figure 3. Model liquids with different surface tension and viscosity [11]. The figure is reproduced with permission from the author. The properties of these fluids are provided in Table 1.
Figure 3. Model liquids with different surface tension and viscosity [11]. The figure is reproduced with permission from the author. The properties of these fluids are provided in Table 1.
Fluids 08 00183 g003
Figure 4. A shadowgraph of the field of view in the imaging plane.
Figure 4. A shadowgraph of the field of view in the imaging plane.
Fluids 08 00183 g004
Figure 5. Illustration of (a) random sampling and (b) sequential sampling techniques.
Figure 5. Illustration of (a) random sampling and (b) sequential sampling techniques.
Fluids 08 00183 g005
Figure 6. A comparison of pure 2D architectures on various filtering intensities and sampling conditions.
Figure 6. A comparison of pure 2D architectures on various filtering intensities and sampling conditions.
Fluids 08 00183 g006
Figure 7. ECOmini-18 vs. ECOmini-C on various coefficients of variation (CV) and sampling conditions.
Figure 7. ECOmini-18 vs. ECOmini-C on various coefficients of variation (CV) and sampling conditions.
Fluids 08 00183 g007
Figure 8. ECOmini-18 vs. ResNet-18 on various coefficients of variation (CV) and sampling conditions.
Figure 8. ECOmini-18 vs. ResNet-18 on various coefficients of variation (CV) and sampling conditions.
Fluids 08 00183 g008
Figure 9. Comparison between random sampling (RS) and sequential sampling (SS) while using (a) ResNet-18 and (b) ECOmini-18.
Figure 9. Comparison between random sampling (RS) and sequential sampling (SS) while using (a) ResNet-18 and (b) ECOmini-18.
Fluids 08 00183 g009
Figure 10. Sequential sampling of eight frames with no offset, a five-frame offset, and a ten-frame offset.
Figure 10. Sequential sampling of eight frames with no offset, a five-frame offset, and a ten-frame offset.
Fluids 08 00183 g010
Figure 11. Performance of ECOmini-18 and ResNet-18 over different coefficient of variation (CV) thresholds with (a) random sampling and (b) sequential sampling strategy.
Figure 11. Performance of ECOmini-18 and ResNet-18 over different coefficient of variation (CV) thresholds with (a) random sampling and (b) sequential sampling strategy.
Fluids 08 00183 g011
Figure 12. Performance of networks when trained on a single or two acquisition batches.
Figure 12. Performance of networks when trained on a single or two acquisition batches.
Fluids 08 00183 g012
Table 1. Test liquids used, viscosity ( η ), density ( ρ ), surface tension ( σ ), Ohnesorge number produced with a 500 μ m nozzle at 25 °C, and components of the liquids.
Table 1. Test liquids used, viscosity ( η ), density ( ρ ), surface tension ( σ ), Ohnesorge number produced with a 500 μ m nozzle at 25 °C, and components of the liquids.
Fluid η ρ σ OhnesorgeComponents
(mPas)(kg/m3)(mN/m)Number (Oh)
A1.099831.97.93 × 10−3Deionized water (DI), Sympatens®
B16.9116965.986.11 × 10−3DI water, Glycerin
C10.5113947.363.98 × 10−3DI water, Glycerin, Myrj® S100®
D1.099870.85.32 × 10−3DI water
E16.9116930.5126.57 × 10−3DI water, Glycerin, Sympatens®
Table 2. Seven deep learning architectures used in this study and the convolution types in the architectures.
Table 2. Seven deep learning architectures used in this study and the convolution types in the architectures.
ArchitectureConvolutionsDescription
ResNet-18
[14]
2D onlyAn 18 layers shallow network containing
convolutions and skip connections
ResNet-C2D onlyA 14 layer variant of ResNet-18
3D-ResNet18
[15]
3D onlyA variant of ResNet-18 containing
purely 3D convolutions
ECO Lite
[7]
2D + 3DContains 2D convolution layers of BNInception
for 2D feature extraction which are
fed to the 3D-ResNet18 architecture
ECO Full
[7]
2D + 3DECO Lite augmented by adding a 2D convolution
stream parallel to the 3D layers
ECOmini-18
[14]
2D + 3DA shallower variant of ECO Full where 2D
convolutions are replaced by layers of ResNet-18
ECOmini-C
[14]
2D + 3DA shallower variant of ECO Full where 2D
convolutions are replaced by layers of ResNet-C
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sardana, P.; Zolfaghari, M.; Miotto, G.; Zengerle, R.; Brox, T.; Koltay, P.; Kartmann, S. DropletAI: Deep Learning-Based Classification of Fluids with Different Ohnesorge Numbers during Non-Contact Dispensing. Fluids 2023, 8, 183. https://doi.org/10.3390/fluids8060183

AMA Style

Sardana P, Zolfaghari M, Miotto G, Zengerle R, Brox T, Koltay P, Kartmann S. DropletAI: Deep Learning-Based Classification of Fluids with Different Ohnesorge Numbers during Non-Contact Dispensing. Fluids. 2023; 8(6):183. https://doi.org/10.3390/fluids8060183

Chicago/Turabian Style

Sardana, Pranshul, Mohammadreza Zolfaghari, Guilherme Miotto, Roland Zengerle, Thomas Brox, Peter Koltay, and Sabrina Kartmann. 2023. "DropletAI: Deep Learning-Based Classification of Fluids with Different Ohnesorge Numbers during Non-Contact Dispensing" Fluids 8, no. 6: 183. https://doi.org/10.3390/fluids8060183

Article Metrics

Back to TopTop