Research article Special Issues

Determinants of the Internet of Things adoption by millennial farmers

  • Indonesia is experiencing difficulties in ensuring the sustainability of the agricultural system as the younger generation experiences reluctance to enter the business of agriculture. Smart farming is believed to be a solution to the difficulty of millennials entering the business of agriculture. One of the main elements of smart farming is the Internet of Things (IoT). This study aims to determine the factors that encourage millennial farmers in Central Java to adopt IoT-based innovations using a behavioral reasoning theory (BRT) perspective. Data were collected from 120 millennial farmers in Central Java; we applied BRT, an analytical technique, to examine IoT adoption by millennial farmers. Primary survey data analysis was carried out by applying structural equation modeling techniques. The results showed that millennial farmers accepting the adoption of IoT technology is a factor of relative advantage and social influence. Meanwhile, the reason for rejecting the adoption of IoT technology is technology anxiety. This research provides information on the reasons for accepting and reasons for rejecting the adoption of IoT in agriculture by millennial farmers in Central Java province, which will be helpful for the government in the design of a program to attract millennials to go into business in agriculture.

    Citation: Mohamad Harisudin, Kusnandar, Erlyna W. Riptanti, Nuning Setyowati, Isti Khomah. Determinants of the Internet of Things adoption by millennial farmers[J]. AIMS Agriculture and Food, 2023, 8(2): 329-342. doi: 10.3934/agrfood.2023018

    Related Papers:

    [1] Qian Shao, Rongrong Xuan, Yutao Wang, Jian Xu, Menglin Ouyang, Caoqian Yin, Wei Jin . Deep learning and radiomics analysis for prediction of placenta invasion based on T2WI. Mathematical Biosciences and Engineering, 2021, 18(5): 6198-6215. doi: 10.3934/mbe.2021310
    [2] Xiaobo Zhang, Donghai Zhai, Yan Yang, Yiling Zhang, Chunlin Wang . A novel semi-supervised multi-view clustering framework for screening Parkinson's disease. Mathematical Biosciences and Engineering, 2020, 17(4): 3395-3411. doi: 10.3934/mbe.2020192
    [3] Keyue Yan, Tengyue Li, João Alexandre Lobo Marques, Juntao Gao, Simon James Fong . A review on multimodal machine learning in medical diagnostics. Mathematical Biosciences and Engineering, 2023, 20(5): 8708-8726. doi: 10.3934/mbe.2023382
    [4] Jing Zhang, Ting Fan, Ding Lang, Yuguang Xu, Hong-an Li, Xuewen Li . Intelligent crowd sensing pickpocketing group identification using remote sensing data for secure smart cities. Mathematical Biosciences and Engineering, 2023, 20(8): 13777-13797. doi: 10.3934/mbe.2023613
    [5] Feng Li, Mingfeng Jiang, Hongzeng Xu, Yi Chen, Feng Chen, Wei Nie, Li Wang . Data governance and Gensini score automatic calculation for coronary angiography with deep-learning-based natural language extraction. Mathematical Biosciences and Engineering, 2024, 21(3): 4085-4103. doi: 10.3934/mbe.2024180
    [6] Jun Yan, Tengsheng Jiang, Junkai Liu, Yaoyao Lu, Shixuan Guan, Haiou Li, Hongjie Wu, Yijie Ding . DNA-binding protein prediction based on deep transfer learning. Mathematical Biosciences and Engineering, 2022, 19(8): 7719-7736. doi: 10.3934/mbe.2022362
    [7] Jia-Gang Qiu, Yi Li, Hao-Qi Liu, Shuang Lin, Lei Pang, Gang Sun, Ying-Zhe Song . Research on motion recognition based on multi-dimensional sensing data and deep learning algorithms. Mathematical Biosciences and Engineering, 2023, 20(8): 14578-14595. doi: 10.3934/mbe.2023652
    [8] Lu Yuan, Yuming Ma, Yihui Liu . Protein secondary structure prediction based on Wasserstein generative adversarial networks and temporal convolutional networks with convolutional block attention modules. Mathematical Biosciences and Engineering, 2023, 20(2): 2203-2218. doi: 10.3934/mbe.2023102
    [9] Tan Gao, Lan Zhao, Xudong Li, Wen Chen . Malware detection based on semi-supervised learning with malware visualization. Mathematical Biosciences and Engineering, 2021, 18(5): 5995-6011. doi: 10.3934/mbe.2021300
    [10] Sakorn Mekruksavanich, Anuchit Jitpattanakul . RNN-based deep learning for physical activity recognition using smartwatch sensors: A case study of simple and complex activity recognition. Mathematical Biosciences and Engineering, 2022, 19(6): 5671-5698. doi: 10.3934/mbe.2022265
  • Indonesia is experiencing difficulties in ensuring the sustainability of the agricultural system as the younger generation experiences reluctance to enter the business of agriculture. Smart farming is believed to be a solution to the difficulty of millennials entering the business of agriculture. One of the main elements of smart farming is the Internet of Things (IoT). This study aims to determine the factors that encourage millennial farmers in Central Java to adopt IoT-based innovations using a behavioral reasoning theory (BRT) perspective. Data were collected from 120 millennial farmers in Central Java; we applied BRT, an analytical technique, to examine IoT adoption by millennial farmers. Primary survey data analysis was carried out by applying structural equation modeling techniques. The results showed that millennial farmers accepting the adoption of IoT technology is a factor of relative advantage and social influence. Meanwhile, the reason for rejecting the adoption of IoT technology is technology anxiety. This research provides information on the reasons for accepting and reasons for rejecting the adoption of IoT in agriculture by millennial farmers in Central Java province, which will be helpful for the government in the design of a program to attract millennials to go into business in agriculture.



    From the birth of the world's first general-purpose computer, "ENIAC" in 1946, people began to explore the possibility of autonomous learning by computers without the limitation of preset modes [1]. Machine learning is a small branch of artificial intelligence, which can predict and infer unknown sample categories or information about new emerging things by learning some regular characteristics of the known data or event forms. When the tasks encountered are very complex or need computer adaptive adjustment, this needs to be done with the help of the powerful data analysis ability of machine learning algorithms [2]. Therefore, it is proposed to investigate whether self-learning can be continued through computers or related algorithms to achieve the purpose of summing up experience, for the system to constantly improve itself. With continuous exploration and the development of technology, the application of computers in various fields is more and more extensive, and artificial intelligence has been rapidly developed to a high level [3]. In daily life, commonly used smart phone applications can provide personalized services according to the user's historical search data; in the medical field, smart phone applications can accumulate experience from medical big data to assist doctors in updating and learning disease treatment methods [4,5,6].

    Machine learning is the core concept in the field of artificial intelligence. There is no clear conclusion about its specific mechanism up to now [7], but it is clear that the main process in machine learning is the selection of a group of representative features to build a model for a learning task, and to achieve self-improvement through continuous learning. Deep learning is a type of machine learning algorithm, which is essentially a multi-layer neural network structure. Compared with the traditional machine learning algorithm, its principle is to increase the network layer to better select data characteristics and improve the ability to handle variability [8]. The calculation process needs to carry out a back propagation (BP) calculation for each layer of the network to achieve the optimal calculation of the objective function, and finally obtain the network model parameters. With the accumulation of experience, how to automatically improve the performance of computer algorithms is the focus of current research. It has been found that the nature of data itself plays an important role in a learning task. If the data is adapted to the learning process, the efficiency of machine learning will be improved and the learning task will be simpler. However, if the data itself has little important information related to the learning task, machine learning may often fail [9,10,11]. From the perspective of biological neuroscience theory, the deep learning model can better simulate the principle of the human brain signal transmission mechanism. Each output neuron is a more deep-seated neural node through weighted calculation and nonlinear transmission, so as to realize the effective flow of original data in the network. Successful feature extraction plays an important role in the field of machine learning, such as in data mining, information classification and retrieval.

    With the explosive growth of multi-source heterogeneous data, big data has become the focus of attention. It is the focus of current research into how to learn efficiently and continuously from massive and complex information. Based on the deep learning model and feature extraction technology, the study analyzes several representative machine learning algorithms, and compares the results of the optimized deep learning model with other algorithms. The research is of great value for extracting representative features from high-dimensional and unlabeled data, thus greatly reducing the amount of labeled data required for learning new tasks.

    Almabdy and Elrefaei (2019) constructed an eight-layer deep neural network model, AlexNet, which reduced the error rate of image recognition to 16.4% and made further breakthroughs in deep learning in the field of image recognition [12]. Basha et al. (2020) invented a deeper VGG deep neural network model, including 16 convolution layers and 3 full connection layers [13]. Wang and Zhang (2020) further analyzed the deep convolution network and constructed DenseNet's deep neural network model, based on the idea of reusing the characteristics of different layers [14]. Du et al. (2018) found that the application of optimization algorithms, such as gradient descent, in the deep learning model can implicitly balance the unbalanced parameters between positive homogeneous layers [15].

    In research into feature extraction, text classification is a challenging field with high-dimensional feature space. Onan et al. (2016) found it very useful to extract keywords of document content and use them as features [16]. On this basis, Onan et al. (2016), based on the prediction performance of the classification algorithm, combined with Bayesian logical regression, and with appropriate weight values assigned to the classifier and each output category, enhanced the prediction performance of emotion classification [17]. Kang (2019) analyzed the basic concepts of machine learning algorithms, obtained the important means to deal with big data at present, and put forward theoretical guidance for the development of the big data era [18]. Fuentes (2018) constructed a deep neural network and completed the verification phase accordingly. Compared with the previous technology, it has better performance and improves the accuracy significantly [19].

    The above research reveals that the algorithm in deep learning has been optimized in continuous improvement, but the deep learning model needs to be further explored in terms of feature extraction from images, speech, and other information. In view of this deficiency, this study discusses several classical machine learning algorithms based on deep learning models and feature extraction. A genetic algorithm is combined with a mutual information feature extraction algorithm in machine learning to form a combined algorithm which improves and optimizes the deep learning model, and realizes an efficient and long-term task learning process.

    Continuous machine learning, also known as Learning to Learn, can not only update the deep learning model online, but also update learning skills according to changing learning progress. Based on this, the related algorithms of machine learning must satisfy the migration of knowledge or structure. It has been concluded that machine learning should possess the following elements: discovering high-level abstract representation of knowledge from raw data; transferring knowledge from completed learning tasks to improve learning ability and complete new learning tasks; possessing the ability to quickly learn new tasks and accumulate and renew knowledge to enhance the level of knowledge; ability to develop knowledge under human guidance and feedback; and to share learning with other systems [20,21]. The structure of a machine learning system is shown in Figure 1.

    Figure 1.  The structure of a machine learning system.

    Knowledge migration: Select and migrate useful knowledge in the knowledge warehouse, which is helpful in promoting the effective learning of new models.

    Knowledge storage: This refers to storing the knowledge that the system learns and which will be used later, playing the function of knowledge warehouse. In terms of knowledge storage, the first method is to store directly the data samples that have been learned, which can preserve the complete data information. However, the high spatial complexity will affect the reading efficiency of the stored information when acquiring knowledge, so the second method is to extract and compress the original information, only preserving the feature representation of the knowledge. The second method is faster than the first method when extracting information, but some detailed data may be omitted.

    Model learning: This mainly involves rapid learning of new tasks and interacting with the knowledge warehouse. It can improve the learning speed of new tasks, while integrating the new learning content with the original knowledge, which makes the knowledge base more abundant. The process of model learning is independent of that of the machine learning system. Knowledge in the knowledge warehouse is not always reflected in the process of learning new models.

    Knowledge integration: The integration of knowledge is a key link in a machine learning system. Its purpose is to ensure that the knowledge warehouse is continuously updated, and to realize the effective transfer of knowledge after learning new tasks. In this process, a machine learning system should screen knowledge, eliminate invalid knowledge and integrate effective knowledge without losing the original key information.

    Guided learning: This process focuses on analyzing the impact of the sequence of learning tasks on the learning performance and efficiency of the system, and emphasizes the importance of learning efficiency. In today's big data environment, because of the huge amount of complex data involved, it will not be able to complete all knowledge learning. By guiding learning, it is possible to arrange learning tasks reasonably and improve the learning efficiency of the system.

    Realization of human-computer interaction and machine-computer interaction: This process is the complement of the whole system, and it can obtain additional relevant information, improve knowledge storage, as well as find errors in the system itself and correct them in a timely manner.

    The ELLA efficient machine learning algorithm was proposed by Paul Ruvolo and Eric Eaton in 2013 on the basis of the GO-MTL (overlapping grouping multi-task) algorithm. The algorithm establishes a prediction model, f(n)(x)=f(n)(x;θ(n)), for each task n by taking θ(n)Rd as the parameter, which can represent the proposed parameter hypothesis through a set of linear combinations of shared bases. When learning a set of tasks simultaneously, a model based on the ELLA algorithm can selectively share information among tasks, such that irrelevant tasks will not affect each other [22]. The structure of the ELLA algorithm is shown in Figure 2. The ELLA algorithm stores a group of shared knowledge bases of potential tasks. When new tasks appear, it helps to learn knowledge from the new tasks by transferring the knowledge learned, and refines the knowledge bases of potential tasks on this basis. Based on this process of maintaining and sharing a knowledge base, the performance of the original learning task model can be effectively improved to meet various requirements of the machine learning algorithm.

    Figure 2.  The structure of the ELLA algorithm.

    The SUPART algorithm was proposed by M. Mishra on the basis of the DG-MTL (non-intersecting multi-task learning) algorithm in 2015. It is also called the Learning to Learn algorithm for supervising task space partitioning [23]. A supervised learning algorithm is used to obtain the segmentation function of task grouping and clustering. When new learning tasks occur, both the segmentation function and the inference function will undergo online updating of model parameters. The objective function of the model based on SUPART algorithm is expressed as:

    R=min1Nnl(ωn,D)+λF(W) (1)

    In the above equation, ωn represents the parameters of task n; λ denotes the normalized parameters of the SUPART model, and F refers to the normal constraints on the whole-task parameter matrix, W. In addition, a segmentation function is introduced to divide the task into A and B regions in space. The corresponding task parameters ωn can be expressed as uA + vn and uB + vn, u represents the common model in the corresponding region, and vn is the private model of the task. Therefore, the objective function of the SUPART model can be improved as follows:

    R=min(1Nn(g=0)lAn+1Nn(g=1)lBn+λF(W)) (2)

    In Eq (2), lAn=1NNil(uA,vn,xi,n,yi,n), lBn=1NNil(uB,vn,xi,n,yi,n). The SUPART model is optimized by a single variable method. After determining the segmentation function, all task parameters are optimized until the model converges iteratively.

    The genetic algorithm was first proposed by Professor Holland of the University of Michigan in the United States, and then summarized to form a simulated evolutionary algorithm [24,25]. The genetic algorithm is an adaptive, probabilistic search algorithm based on Darwin's theory of biological natural selection and evolution. It generates new populations through genetic operations such as selection, crossover and mutation, and then realizes the gradual evolution of populations.

    The genetic algorithm takes the coding of decision variables as the object of operation. In the process of optimization and calculation, it can imitate the genetic evolution mechanism of organisms in nature, search for groups composed of multiple groups, and then carry out selection, crossover and mutation operations to generate new groups containing distinct sets of group information. With the continuous evolution of the population, the new population will produce more high-quality individuals. The main components of a genetic algorithm are:

    1) coding: the common methods are binary coding and floating-point coding, which refer to the process of describing the feasible solution of the problem;

    2) population initialization: as a random search method, the common method of population initialization is: random generation of a population containing multiple individuals, with reasonable individuals generated by using algorithms or a priori knowledge to avoid too close similarity between individuals in initialization, so as to converge the evolutionary process;

    3) fitness function: mainly converting the objective function of the problem into a fitness function, and then determining the selection probability of the individual, which is also the core of the genetic algorithm;

    4) selection, mutation and intersection. A crossover operator has strong global search ability, so, as the main operation, the mutation operator is an auxiliary operation. The basic step diagram of a genetic algorithm is shown in Figure 3.

    Figure 3.  Steps of a genetic algorithm.

    In Wrapper feature selection, based on a genetic algorithm, each individual of the genetic algorithm represents a subset of features. When the population is initialized, a group of different individuals will be generated randomly. The fitness function includes two parts: the classification accuracy of the classifier and the size of the feature subset. The higher the classification accuracy of the classifier, the smaller the dimensions of the feature subset can be; the larger the fitness value of an individual is, the more chance of its features being inherited by the next generation. The genetic algorithm has low dependence on the problem of evaluating individuals by the fitness function. Moreover, the genetic algorithm has strong global search ability, and is widely used in various fields where feature selection is needed. However, when feature analysis is needed for large-scale data, the Wrapper structure based on a genetic algorithm is inefficient and will not generally be applied.

    With the deepening of information theory and the rapid development of information engineering, the field that it can cover goes far beyond the narrow sense of communication engineering. It has been combined with artificial intelligence, automatic control, system engineering, and other disciplines, and developed into a comprehensive discipline-Information Science.

    In information theory, the general source is represented by X, and information entropy is represented by H(X), which is a number representing the overall information measure of the source in an average sense. The information entropy can represent the impurity of X. From the point of view of information theory, if the source is represented by X and the destination is represented by Y, then H(X) is a measure of the prior uncertainty of input variable X before it receives Y, so it is also called a priori entropy; H(X | Y) represents the average uncertainty of input variable X after receiving output Y, which is mainly caused by noise. H (X | Y) is called channel ambiguity. It can be seen that, after receiving Y, some information is obtained and some uncertainties are eliminated. In information theory, the mutual information between X and Y is expressed by I(X; Y), which can be expressed as:

    I(X;Y)=H(X)H(X|Y) (3)

    I(X; Y) can reflect the interdependence between input and output random variables and the degree of statistical constraints between them. Generally, conditional entropy H(X | Y) ≤ unconditional entropy H(X), so mutual information I(X; Y) will take a non-negative value. It should be noted that mutual information I(X; Y) and I(Y; X) are equal because of the symmetry of mutual information, which is very valuable for measuring the interdependence between the two attributes.

    Mutual information-based feature prescreening is actually a category of feature selection methods based on correlation. The main route of feature selection based on correlation is to evaluate the correlation between features and category labels on the basis of a specific definition of correlation. After sorting the features in descending order, a feature subset which correlates highly with category labels is selected to achieve the goal of dimensionality reduction. This selection method is very important for feature selection. Besides mutual information, there are linear correlation coefficients to measure the correlation of two attributes. The linear correlation coefficients of X and Y attributes are expressed as follows:

    r=i(xi¯x)(yi¯y)i(xi¯x)2i(yi¯y)2 (4)

    In the equation, ¯x and ¯y indicate the mean value of X and Y, respectively, and the range of r is -1 ~ 1. If X and Y are completely linear, then r is -1 or 1; if X and Y are not linear, then r is 0. The linear correlation coefficient is limited to the calculation of the correlation between numerical attributes, and it cannot capture non-linear correlation between attributes.

    The second stage of the combinatorial feature selection algorithm (MI-GA) is to search the final feature subset through a genetic algorithm in the feature set after mutual information screening. Using binary coding, the length of coding is the number of features in the subset after the output of the algorithm. The result of feature ranking based on mutual information is used to guide the population initialization, and the result of feature ranking based on mutual information provides the prior knowledge of the problem for the genetic algorithm.

    A deep learning model belongs to a kind of non-linear machine learning model. Because its concept is inspired by neural networks in the biological brain, it is also called a deep neural network model. Neuron cells are the basic units of the brain's biological mechanism. Each neuron cell is differentiated into dendrites and axons upstream and downstream, respectively, as shown in Figure 4. When the nerve signal starts to transmit under stimulation, the upstream neurons release chemicals, and the dendrites receive the signals and process them, before the axons release them to the downstream neurons. Based on this structure of the brain's neural network, researchers constructed an artificial neuron structure, as shown in Figure 5. The artificial neuron can receive more than one signal input, and then a nonlinear activation function can be driven by weighted summation results.

    Figure 4.  Neuronal cell structure.
    Figure 5.  Cell structure of artificial neurons.

    A convolutional neural network (CNN) in deep learning is a kind of neural network that can deal with time series data or image data and other data with a grid structure. In other words, at least one layer of the network uses a convolution operation to replace the general matrix multiplication operation. Time series data can be regarded as one-dimensional data on the time axis; image data is two-dimensional data with two dimensions (height and width); and video data can be regarded as three-dimensional data. Compared with the traditional full-connection layer, a CNN is characterized by small parameters and more efficient calculation.

    A recurrent neural network (RNN) is a kind of neural network algorithm. It can be seen as a process of generating a directed graph by connecting the edges of different nodes. This structure enables it to deal dynamically with different behaviors in a time series. Compared with the traditional feedforward neural network, this network can use the internal state to process the input of the sequence. In addition, the structure also enables it to be applied to tasks such as handwriting recognition and speech recognition.

    Reinforcement learning is a widely used machine learning algorithm in complex scenes. This algorithm has been widely used in machine control tasks. In recent years, with the rapid development of deep learning, the reinforcement learning method has been widely used in computer vision, automatic driving, and other new fields. Recently, there are many reinforcement learning algorithms based on deep learning. In 2015, a deep Q network (DQN) was born, in which the Q learning method was used to solve the action value function problem in reinforcement learning. Unlike the traditional manual feature extraction method, the model can carry out end-to-end learning.

    At present, most machine learning models are unable to cope with inconsistent input distribution and high dimensionality. Therefore, an algorithm is proposed that combines a deep learning model with lifelong machine learning, based on the ELLA algorithm; on this basis, a new deep learning model (HLLA, Hierarchical lifelong learning algorithm) is ultimately formed. By adding the feature extraction model of unsupervised deep learning, the information data after feature extraction can better fit the same distribution, which is helpful for knowledge transfer of different distributions in a lifelong machine learning system, while input feature representation can improve the performance of traditional supervised machine learning models when the number of tags is small and the dimensionality of data is high.

    The process of machine learning based on deep learning model optimization is divided into four main steps: unsupervised pretraining; maintenance and updating of the feature model; feature extraction; and supervised machine learning. This deep learning model has the following advantages:

    1) the deep learning model has been proved to be effective in preventing the subsequent classification model falling into a local optimal solution under the condition of insufficient labels, which can greatly improve the efficiency of machine learning.

    2) the deep learning model generates neural network weights in the unsupervised pre-training process, which will be a significant leap forward in machine learning in the context of big data.

    3) a hierarchical feature extraction network can increase the complexity of the model and greatly improve the performance of the learning model in dealing with non-linear separable data and high-dimensional data.

    4) data will contain the features trained and learned by many unlabeled data before entering the deep learning model, which is similar to the process of knowledge transfer and has a significant effect on subsequent integration and refinement of knowledge.

    The HLLA algorithm based on DBN and ELLA assumes that there are a total of Tmax supervised learning tasks Z(1), Z(2), Z(Tmax()), and many unlabeled data xu, in accordance with the distribution of learned tasks. The label of each learning task Z(t)=(X(t),y(t)) is determined by a real implicit function f(t). At any time point, the learning system receives a batch of labeled samples from task t, which may be from a new task or a task that has been learned. T represents the total number of tasks learned before. After the training data are received, the learning system aims to learn the learning model f(t) of each task through the given training samples to approach the real objective function f(t). Supposing that the DBN model has nL layers, the weight connection network of the l-th layer is Wl; the bias vector is bl; the number of hidden neurons in each layer is dl; the response function of neurons is the sigmoid function; and the objective function of the HLLA model can be expressed as follows:

    eT(L,Wl)=1Ttmins(t){1ntnTiloss(f(h(t)i,nL;Ls(t)),y(t)i)+μs(t)1}+λL2F (5)

    h(t)i,nL shows the feature representation of the i-th input of the task t in the l-th layer, and ntrepresents the number of training samples of task t. The parameters θ(t) of each task model are composed of column vectors in matrix L by linear combination. The HLLA model has two main optimization objectives: knowledge base matrix L and network matrix Wl for layer connection of the depth model. If L is optimized first and then Wl, the characteristics of the last layer will be changed due to the change of network weights. For each new task, the deep learning feature network Wl needs to be optimized first, and then the knowledge base matrix L is optimized.

    Before the learning process starts, it effectively uses information from several unlabeled data Xu, and constructs a DBN feature extraction model through unsupervised training, so that the labeled training data ((X(t),y(t))) is more consistent with a distribution after feature extraction, making the initial value of the ELLA knowledge base better represent and fit the distribution of task parameters, as well as enabling the subsequent training process to achieve convergence more easily.

    The HLLA model consists of a fully shared feature extraction model based on deep learning and supervised lifelong machine learning. The learning process of the new model includes four main steps:

    1) before accepting the continuous learning task, sample extensive task-related unlabeled data from the spatial distribution of the input data, and then train the DBN network model by using the unsupervised reconstruction pretraining method in the DBN model training step described in the previous section;

    2) fine-tune and update the generated DBN neural network to adjust the feature network to avoid affecting the representation of features by distribution drift;

    3) extract the layer by layer features of the current DBN network pair by a task that is about to enter the lifelong learning system;

    4) supervised lifelong machine learning (i.e., the process of parameter sharing and hypothesis integration between tasks).

    In order to bridge the gap between low-level visual features and high-level semantics, a CNN is applied to image feature extraction. In addition, by introducing the migration learning idea and using the pre-training model obtained from the ImageNet data set, the problem of model over-fitting caused by the initial training can be solved. A sample image is used to adjust the model parameters through BP, so as to generate a new model suitable for the image classification task.

    The AlexNet network has the characteristics of local connection, weight sharing, pooling operation and so on. As an example of a CNN, AlexNet is composed of five convolution layers, three pooling layers, and three full-connection layers, including 60 million parameters and 650,000 neurons, which can gradually transform low-level features into high-level features. Among them, the first several layers mainly obtain low-level features, such as color, edge, texture, and so on, while the second several layers can produce high-level features for direct recognition and classification.

    Transfer learning can transfer knowledge from the original domain to the target task. In order to improve the generalization ability of AlexNet to the sample image, the activation function of the training CNN is selected as the rectification linear unit to avoid gradient disappearance during the training process of the model; local response normalization is used to smooth the feature image, and multi-layer output feature combination can improve the image classification accuracy.

    Application in an image classification and recognition task: in order to evaluate the proposed model, this experiment uses a large-scale image description data set, MSCOCO, including 81,903 training images, 39,893 verification images, and 40,015 test images. Finally, 10,000 samples are extracted from the training set as the development set, and the performance of unsupervised feature learning is evaluated.

    First, many small patches are randomly extracted from the original training image on MSCOCO. Each patch is a 6 × 6-pixel image, which is represented by 36-dimensional vector RD. A total of 40,000 patches are extracted and used for unsupervised feature learning. The classification error rate of the linear support vector machine (SVM) classifier, based on unsupervised learning and feature training in the MSCOCO database, is shown in Table 1.

    Table 1.  Error rate of linear SVM classifier based on unsupervised learning and feature training in MSCOCO database.
    Model/K 400 800 1200 1600
    Kmeans 1.40 1.30 1.15 1.12
    Spkmeans 1.08 0.90 0.87 0.83
    movMF 0.90 0.81 0.82 0.83
    PCA-movMF 0.87 0.76 0.74 0.75
    Optimization model 0.75 0.70 0.63 0.67

     | Show Table
    DownLoad: CSV

    Each feature extractor performs a convolution operation on the image to get the feature expression of each local patch of the image. Then, the image is divided into four quadrants, and each quadrant's feature gets a feature vector by accumulating. In order to verify the quality of the feature expression extracted by different feature extractors, a simple linear SVM is trained in the back-end to classify the image by taking the extracted feature vector as the input, and the quality of the extracted feature expression is judged by the accuracy of the image classification. In the case of the same back-end classifier, the lower the error rate of image classification, the better the feature expression extracted by the corresponding front-end feature extractor. The experimental results show that the deep learning model based on feature extraction can further improve the performance of image classification. When K = 1200, it can achieve a classification error rate of 0.63%, which reflects an excellent performance by the unsupervised database algorithm based on the model.

    Application in speech recognition task: a speech recognition database is selected to verify the impact of various algorithms on the performance evaluation of data sets. A total of 90 subjects' pronunciation of 26 English letters is collected. All subjects are divided into three groups according to the similarity of pronunciation: A, B and C, 30 people in each group. Sample feature extraction includes information such as the spectrum coefficient feature and the acoustics feature, with a total of 367 dimensions. An overview of the experimental voice database is shown in Table 2.

    Table 2.  Experimental speech database information.
    Name Number of tasks Dimensions Total sample size
    Speech recognition 3 367 4019

     | Show Table
    DownLoad: CSV

    By comparing the STL algorithm, DG-MTL algorithm, ELLA algorithm, and HLLA algorithm, the reliability of the deep model based on the HLLA algorithm was analyzed. The experimental results show that the HLLA algorithm optimized in this paper performed best in all cases on the speech recognition database. Especially at the key points of 50 and 100%, its classification accuracy is higher than that of other classification algorithms. The classification accuracy of the four algorithms in the speech recognition database is shown in Table 3, and the comparison of the classification accuracy of each algorithm in the speech database is shown in Figure 6.

    Table 3.  Classification accuracy (%) of four algorithms in speech recognition database.
    Algorithm 10% 20% 30% 40% 50% 60% 70% 80% 90% 100%
    STL 65.5 71.2 74.7 76.8 78.3 80.3 82.0 82.4 82.8 83.6
    DG-MTL 67.8 73.6 80.3 83.3 84.8 84.5 84.9 85.7 86.2 85.9
    ELLA 73.1 77.3 82.2 86.0 87.6 89.4 91.3 92.8 93.3 94.0
    HLLA 79.1 80.5 84.6 87.9 89.8 90.9 93.0 93.7 94.8 95.5

     | Show Table
    DownLoad: CSV
    Figure 6.  Comparison of classification accuracy of speech database based on different algorithms.

    According to the experimental results of four different training set sizes in the speech recognition multi-task database, the abscissa in Figure 6 is the proportion of the training set used (100% indicates using all training data), and the ordinate is the classification accuracy (ACC) index of various algorithms in various situations; the higher the value, the better the performance. The HLLA algorithm proposed performs the best in all cases on this database. Even when the number of tags in each task is reduced, the new HLLA algorithm still maintains good performance because the features extracted from feature extraction carry more structured information than the original data and are more consistent with one distribution. On the other hand, the performance of the ELLA algorithm alone is seriously affected because the amount of information is not enough to fit the overall distribution.

    Figure 7 describes the task correlation matrix learned by the HLLA algorithm on two databases. The abscissa and ordinate of the matrix represent the task number, and the depth of the grid colour is proportional to the correlation of the corresponding coordinate task. The Animals dataset contains 50 categories and 30,475 images, each of which has only one category label. Two kinds of dogs and four kinds of birds are selected as the normal group, and the remaining animals are randomly selected as the parent group, comprising the two-class multi-task database. According to the learning of task correlation based on reconstruction error of the HLLA algorithm, in the Landmine database, the first 14 tasks are highly correlated, belonging to the first group, while tasks 15–30 are highly-correlated, belonging to the second group, and the correlation between the two groups is low. This learning result is also consistent with the actual situation that the database itself comes from two different surface areas. On the Animals database, the HLLA algorithm also successfully divided bird and dog into three independent groups, dog 1, bird 1 ~ bird 4, and dog 2. The statistical results of animal image classification accuracy in Figure 8 reveal that, the HLLA algorithm model is the optimal algorithm in almost all cases, especially in the case of fewer tags, there is a considerable gap between HLLA and other algorithms.

    Figure 7.  Task correlation matrix learned by HLLA algorithm.
    Figure 8.  Accuracy of animal image classification.

    After the ELLA and HLLA algorithms are learned, a deep learning model based on ELLA is constructed and optimized, so that the learning equipment can carry out more lasting learning, and can learn from the huge and complex data centralized databases and groups in the big data era. The relevant information and research results can improve the learning equipment for learning. It can provide the design concept for the new learning equipment combined with learning technology in the future, which is a major achievement to open up practice and connect the future. The great success of deep learning depends on its powerful ability to fit data, the explosive growth of data available for training, and the development of distributed optimization algorithms. Deep learning has brought unexpected success in many complex tasks. The main concept of machine learning is to select a group of representative features to build a model for a learning task, and to achieve self-improvement through continuous learning. In recent years, many machine learning algorithms have been shown to be negatively affected by irrelevant and redundant features. Therefore, selecting a representative feature subset from the original feature set is of positive significance to many problems in the field of machine learning. With the explosive growth of multi-source heterogeneous data, big data has become the focus of attention. How to learn efficiently and continuously in huge and complex information is the focus of current research. The existing trend in lifelong machine learning algorithm for high-dimensional data is to use a traditional dimensionality reduction method (such as principal component analysis) to reduce the data dimensionality first and then input it into the lifelong machine learning system. This method, however, does not fully utilize the data information and knowledge of previous tasks, so the effect of some high-dimensional multi-task databases is not ideal. On the other hand, if the dimensionality is not reduced, the multi-task high-dimensional database can be directly input into the existing lifelong machine learning system, which will lead to excessively long system learning time, so that the system does not have practicability.

    Based on the deep learning model and feature extraction, several representative machine learning algorithms are analyzed, and the results of the optimized deep learning model in practical applications are compared with those of other algorithms. First, the basic theory of machine learning and deep learning is introduced, and the structure of a machine learning system for dealing with multiple tasks in different distributions is simply analyzed. Based on the model structure of a deep neural network, a more effective acoustic model and language model of speech recognition is studied. How to further optimize the structure of the existing deep neural network to improve the performance of the model is discussed, aiming to explore the working mechanism of the deep neural network. In addition, a deep learning model for feature information sharing is proposed, based on the combined feature selection algorithm of mutual information and a genetic algorithm. A feature learning network can be a kind of deep learning feature extraction model which can update network parameters by using a back propagation algorithm. A deep sharing network can extract highly representative features from high-dimensional and unlabeled data, thus greatly reducing the amount of tag data needed for learning new tasks.

    The experiments show that the optimized HLLA algorithm performs best in all cases of the speech recognition database, and the classification accuracy is significantly higher than that of other algorithm models. The deep learning model based on feature extraction can further improve the performance of image classification. When K = 1200, it can achieve a classification error rate of 0.63%, which reflects excellent performance of the unsupervised database algorithm based on the model. In the model experiment stage, the extraction of subjects' voices is carried out in a relatively quiet environment, without considering the interference of noise. In a practical application, it is easy to be disturbed by the external environment, which leads to inaccurate information extraction, which will be the part to be optimized in the follow-up study.

    The study focuses on a feature selection algorithm for the data in large-scale integrated deep learning according to features, but it is limited to the joint update sequence of vector combination in feature framework and knowledge database. When the vector groups in the feature framework and knowledge base are updated separately and sequentially, the change of the feature network may no longer be suitable for distributed learning tasks. In the future research, the focus will be how to carry out effective feature selection and feature network division in the unsupervised learning field, to avoid the negative transfer problem in the grouping of related learning tasks. Moreover, in the integrated deep learning based on feature selection, the proposed algorithm does not deliberately increase the difference degree of a classifier. In the feature selection of integrated deep learning, the focus of future research will be how to improve the accuracy of a single classifier, thus further increasing the differences between them.

    We would like to thank you for following the instructions above very closely in advance.

    The authors declare no conflict of interest.



    [1] Haryanto Y, Effendy L, Tri Yunandar D (2022) Characteristics of millenial farmers in rice center area in West Java. J Penyul 18: 25–35. https://doi.org/10.25015/18202236982 doi: 10.25015/18202236982
    [2] Nugroho AD, Waluyati LR, Jamhari J (2018) Efforts of engage youth generation to working on agricultural sector in Yogyakarta Province. JPPUMA J Ilmu Pemerintah dan Sos Polit Univ Medan Area 6: 76–95. https://doi.org/10.31289/jppuma.v6i1.1252 doi: 10.31289/jppuma.v6i1.1252
    [3] Arditia MR, Lubis JF, Saragih B, et al. (2021) Millenial work behaviour and it's impact to office design. IOP Conf Ser Earth Environ Sci 794: 012179. https://doi.org/10.1088/1755-1315/794/1/012179 doi: 10.1088/1755-1315/794/1/012179
    [4] Hamdani C (2020) Factors affecting the performance of millennial farmer farming alumni of entrepreneurship training for young farmers in central Java Province (Faktor yang Mempengaruhi Kinerja Berusaha Tani Petani Milenial Alumni Pelatihan Kewirausahaan Bagi Petani Muda di Provinsi Jawa Tengah). J Agriwidya 1: 61–73.
    [5] Gagliardi G, Cosma AIM, Marasco F (2022) A decision support system for sustainable agriculture: The case study of coconut oil extraction process. Agronomy 12: 177. https://doi.org/10.3390/agronomy12010177 doi: 10.3390/agronomy12010177
    [6] Wolfert S, Ge L, Verdouw C, et al. (2017) Big data in smart farming—A review. Agric Syst 153: 69–80. https://doi.org/10.1016/j.agsy.2017.01.023 doi: 10.1016/j.agsy.2017.01.023
    [7] Dryancour G (2017) Smart agriculture for all farms. CEMA; Eur Agric Mach Ind Assoc 32: 1–23.
    [8] Zhai Z, Martínez JF, Beltran V, et al. (2020) Decision support systems for agriculture 4.0: Survey and challenges. Comput Electron Agric 170: 105256. https://doi.org/10.1016/j.compag.2020.105256 doi: 10.1016/j.compag.2020.105256
    [9] De Clercq M, Vatz A, Biel A (2018) Agriculture 4.0: The Future of Farming Technology.
    [10] Ronaghi MH, Forouharfar A (2020) A contextualized study of the usage of the Internet of things (IoTs) in smart farming in a typical Middle Eastern country within the context of Unified Theory of Acceptance and Use of Technology model (UTAUT). Technol Soc 63: 101415. https://doi.org/10.1016/j.techsoc.2020.101415 doi: 10.1016/j.techsoc.2020.101415
    [11] Mukti GW, Budi Kusumo RA, Qanti SR (2017) Successful behavior of young entrepreneurial farmers graduated from the faculty of agriculture, Padjadjaran University (Perilaku Sukses Petani Muda Wirausaha Lulusan Fakultas Pertanian Universitas Padjadjaran). J Agribisnis Terpadu 10: 221–234. https://doi.org/10.33512/jat.v10i2.5076 doi: 10.33512/jat.v10i2.5076
    [12] Nargotra M, Khurjekar MJ (2020) Green house based on IoT and AI for societal benefit, 2020 International Conference on Emerging Smart Computing and Informatics, 109–112. https://doi.org/10.1109/ESCI48226.2020.9167637
    [13] Zhai Z, Martínez JF, Beltran V, et al. (2020) Decision support systems for agriculture 4.0: Survey and challenges. Comput Electron Agric 170: 105256. https://doi.org/10.1016/j.compag.2020.105256 doi: 10.1016/j.compag.2020.105256
    [14] Liakos KG, Busato P, Moshou D, et al. (2018) Machine learning in agriculture: A review. Sensors (Switzerland) 18: 2674. https://doi.org/10.3390/s18082674 doi: 10.3390/s18082674
    [15] Muangprathub J, Boonnam N, Kajornkasirat S, et al. (2019) IoT and agriculture data analysis for smart farm. Comput Electron Agric 156: 467–474. https://doi.org/10.1016/j.compag.2018.12.011 doi: 10.1016/j.compag.2018.12.011
    [16] Ihsaniyati H, Setyowati N, Pardono (2022) Factors motivating the adoption of geographical indication-based quality standards among Robusta coffee farmers in Indonesia. Int J Bus Soc 23: 207–225. https://doi.org/https://doi.org/10.33736/ijbs.4609.20221 doi: 10.33736/ijbs.4609.20221
    [17] Hsu TC (2005) Research methods and data analysis procedures used by educational researchers. Int J Res Method Educ 28: 109–133. https://doi.org/10.1080/01406720500256194 doi: 10.1080/01406720500256194
    [18] Blakeslee JR (2020) Effects of high-fidelity simulation on the critical thinking skills of baccalaureate nursing students: A causal-comparative research study. Nurse Educ Today 92: 104494. https://doi.org/10.1016/j.nedt.2020.104494 doi: 10.1016/j.nedt.2020.104494
    [19] Nurrizky M, Harisudin M, Barokah U (2023) Influence of experiential marketing to consumer satisfaction and repurchase intentions : "Goreng" Taichan restaurants as a case study. Int J Sustain Dev Plan 18: 247–253. https://doi.org/https://doi.org/10.18280/ijsdp.180126 doi: 10.18280/ijsdp.180126
    [20] Banendro S (2019) Millennial Farmers Boost Central Java's Agricultural Export Value(Petani Milenial Dongkrak Nilai Ekspor Pertanian Jawa Tengah). https://humas.jatengprov.go.id/detail_berita_gubernur?id = 3692
    [21] Kim M (2021) Conceptualization of e-servicescapes in the fitness applications and wearable devices context: Multi-dimensions, consumer satisfaction, and behavioral intention. J Retail Consum Serv 61: 102562. https://doi.org/10.1016/j.jretconser.2021.102562 doi: 10.1016/j.jretconser.2021.102562
    [22] Cahyadinata I, Nusril, Gushevinalti (2020) Descriptive, correlation analysis and analytical hierarchy process of coastal community empowerment of Bengkulu City, Indonesia. Int J Adv Sci Eng Inf Technol 10: 1304–1310. https://doi.org/10.18517/ijaseit.10.3.2659 doi: 10.18517/ijaseit.10.3.2659
    [23] Vrontis D, Chaudhuri R, Chatterjee S (2022) Adoption of digital technologies by SMEs for sustainability and value creation : Moderating role of entrepreneurial orientation. Sustainability 14: 7949. https://doi.org/10.3390/su14137949 doi: 10.3390/su14137949
    [24] Hair JF, Hult GTM, Ringle CM, et al. (2014) A primer on partial least squares structural equation modeling (PLS-SEM). Eur J Tour Res 6: 211–213.
    [25] Elijah O, Rahman TA, Orikumhi I, et al. (2018) An overview of Internet of Things (IoT) and data analytics in agriculture: Benefits and challenges. IEEE Internet Things J 5: 3758–3773. https://doi.org/10.1109/JIOT.2018.2844296 doi: 10.1109/JIOT.2018.2844296
    [26] Coutu A, Mottelet S, Guérin S, et al. (2022) Methane yield optimization using mix response design and bootstrapping: application to solid-state anaerobic co-digestion process of cattle manure and damp grass. Bioresour Technol Reports 17: 100883. https://doi.org/10.1016/j.biteb.2021.100883 doi: 10.1016/j.biteb.2021.100883
    [27] Ammad S, Alaloul WS, Saad S, et al. (2021) Personal Protective Equipment (PPE) usage in construction projects: A systematic review and smart PLS approach. Ain Shams Eng J 12: 3495–3507. https://doi.org/10.1016/j.asej.2021.04.001 doi: 10.1016/j.asej.2021.04.001
    [28] Walter A, Finger R, Huber R, et al. (2017) Smart farming is key to developing sustainable agriculture. Proc Natl Acad Sci U S A 114: 6148–6150. https://doi.org/10.1073/pnas.1707462114 doi: 10.1073/pnas.1707462114
    [29] Sekaran U, Bougie R (2016) Research Methods for Business: A Skill- Building Approach, Wiley & Sons, West Sussex.
    [30] Pillai R, Sivathanu B (2020) Adoption of internet of things (IoT) in the agriculture industry deploying the BRT framework. Benchmarking 27: 1341–1368. https://doi.org/10.1108/BIJ-08-2019-0361 doi: 10.1108/BIJ-08-2019-0361
    [31] Tandon A, Dhir A, Kaur P, et al. (2020) Behavioral reasoning perspectives on organic food purchase. Appetite 154: 104786. https://doi.org/10.1016/j.appet.2020.104786 doi: 10.1016/j.appet.2020.104786
    [32] Rachmawati RR (2020) Smart farming 4.0 to build advanced, independent, and modern Indonesian agriculture. Forum Penelit Agro Ekon 38: 137–154.
    [33] Westaby JD, Probst TM, Lee BC (2010) Leadership decision-making: A behavioral reasoning theory analysis. Leadersh Q 21: 481–495. https://doi.org/10.1016/j.leaqua.2010.03.011 doi: 10.1016/j.leaqua.2010.03.011
    [34] Westaby JD (2005) Comparing attribute importance and reason methods for understanding behavior: An application to internet job searching. Appl Psychol 54: 568–583. https://doi.org/10.1111/j.1464-0597.2005.00231.x doi: 10.1111/j.1464-0597.2005.00231.x
    [35] Akter S, D'Ambra J, Ray P (2011) An evaluation of PLS based complex models: The roles of power analysis, predictive relevance and GoF index, AMCIS 2011 PROCEEDINGS, 1–7.
    [36] Spielhofer R, Thrash T, Hayek UW, et al. (2021) Physiological and behavioral reactions to renewable energy systems in various landscape types. Renew Sustain Energy Rev 135: 21–24. https://doi.org/10.1016/j.rser.2020.110410 doi: 10.1016/j.rser.2020.110410
    [37] Mandhani J, Nayak JK, Parida M (2020) Interrelationships among service quality factors of Metro Rail Transit System: An integrated Bayesian networks and PLS-SEM approach. Transp Res Part A Policy Pract 140: 320–336. https://doi.org/10.1016/j.tra.2020.08.014 doi: 10.1016/j.tra.2020.08.014
    [38] Harisudin M, Adi RK, Qonita RRA (2022) Synergy Grand Strategy Matrix, Swot and Qspm as determinants of Tempeh product development strategy. J Sustain Sci Manag 17: 62–82. https://doi.org/10.46754/jssm.2022.08.004 doi: 10.46754/jssm.2022.08.004
    [39] Ellitan L (2022) Increasing repurchase intention through experiential marketing and customer satisfaction. ULIL ALBAB J Ilm Multidisiplin 1: 3559–3565.
    [40] Pakura S, Rudeloff C (2020) How entrepreneurs build brands and reputation with social media PR: Empirical insights from start-ups in Germany. J Small Bus Entrep 35: 153–180. https://doi.org/10.1080/08276331.2020.1728490 doi: 10.1080/08276331.2020.1728490
    [41] Costanza DP, Ravid DM, Slaughter AJ (2021) A distributional approach to understanding generational differences: What do you mean they vary? J Vocat Behav 127: 103585. https://doi.org/10.1016/j.jvb.2021.103585 doi: 10.1016/j.jvb.2021.103585
    [42] Gupta A, Arora N (2017) Consumer adoption of m-banking: A behavioral reasoning theory perspective. Int J Bank Mark 35: 733–747. https://doi.org/10.1108/IJBM-11-2016-0162 doi: 10.1108/IJBM-11-2016-0162
    [43] Munir R, Beh LS (2019) Measuring and enhancing organisational creative climate, knowledge sharing, and innovative work behavior in startups development. Bottom Line 32: 269–289. https://doi.org/10.1108/BL-03-2019-0076 doi: 10.1108/BL-03-2019-0076
    [44] Lombardi M, Pascale F, Santaniello D (2021) Internet of things: A general overview between architectures, protocols and applications. Inf 12: 87. https://doi.org/10.3390/info12020087 doi: 10.3390/info12020087
    [45] Tanveer A, Zeng S, Irfan M, et al. (2021) Do perceived risk, perception of self-efficacy, and openness to technology matter for solar pv adoption? An application of the extended theory of planned behavior. Energies 14:5008. https://doi.org/10.3390/en14165008 doi: 10.3390/en14165008
    [46] Gursoy D, Maier TA, Chi CG (2008) Generational differences: An examination of work values and generational gaps in the hospitality workforce. Int J Hosp Manag 27: 448–458. https://doi.org/10.1016/j.ijhm.2007.11.002 doi: 10.1016/j.ijhm.2007.11.002
    [47] Westaby JD (2005) Behavioral reasoning theory: Identifying new linkages underlying intentions and behavior. Organ Behav Hum Decis Process 98: 97–120. https://doi.org/10.1016/j.obhdp.2005.07.003 doi: 10.1016/j.obhdp.2005.07.003
  • This article has been cited by:

    1. Xiaoyan Gao, Ning Cao, Development Model of College English Education Service Talents Based on Nonlinear Random Matrix, 2022, 2022, 1563-5147, 1, 10.1155/2022/2725506
    2. Muhammad Bilal Shoaib Khan, Muhammad Saqib Nawaz, Rashad Ahmed, Muhammad Adnan Khan, Amir Mosavi, Intelligent breast cancer diagnostic system empowered by deep extreme gradient descent optimization, 2022, 19, 1551-0018, 7978, 10.3934/mbe.2022373
    3. Khalil Aljohani, Turki Turki, Automatic Classification of Melanoma Skin Cancer with Deep Convolutional Neural Networks, 2022, 3, 2673-2688, 512, 10.3390/ai3020029
    4. Yonghong Zhang, Donglin Xie, Wei Tian, Huajun Zhao, Sutong Geng, Huanyu Lu, Guangyi Ma, Jie Huang, Kenny Thiam Choy Lim Kam Sian, Construction of an Integrated Drought Monitoring Model Based on Deep Learning Algorithms, 2023, 15, 2072-4292, 667, 10.3390/rs15030667
    5. Nie Chen, Shahid Mumtaz, Research on E-Commerce Database Marketing Based on Machine Learning Algorithm, 2022, 2022, 1687-5273, 1, 10.1155/2022/7973446
    6. Youping Wu, Guoqiang Tao, Zaoli Yang, Application of a New Loss Function-Based Support Vector Machine Algorithm in Quality Control of Measurement Observation Data, 2022, 2022, 1563-5147, 1, 10.1155/2022/7266719
    7. Chengxin Li, Tingzhen Zhang, Jun Li, Identifying autism spectrum disorder in resting-state fNIRS signals based on multiscale entropy and a two-branch deep learning network, 2023, 383, 01650270, 109732, 10.1016/j.jneumeth.2022.109732
    8. Min Zhao, Jin Wan, Wenzhi Qin, Xin Huang, Guangdi Chen, Xinyuan Zhao, A machine learning-based diagnosis modelling of type 2 diabetes mellitus with environmental metal exposure, 2023, 01692607, 107537, 10.1016/j.cmpb.2023.107537
    9. Suyao Wei, Zhihui Wu, The Application of Wearable Sensors and Machine Learning Algorithms in Rehabilitation Training: A Systematic Review, 2023, 23, 1424-8220, 7667, 10.3390/s23187667
    10. Jiamin Huang, Yang Gao, Yanhong Chang, Jiajie Peng, Yadong Yu, Bin Wang, Machine Learning in Bioelectrocatalysis, 2024, 11, 2198-3844, 10.1002/advs.202306583
    11. Ruolin Du, Ruiqi Xin, Han Wang, Wenkai Zhu, Rui Li, Wei Liu, Machine learning: An accelerator for the exploration and application of advanced metal-organic frameworks, 2024, 490, 13858947, 151828, 10.1016/j.cej.2024.151828
    12. Yang Meng, Qiang Xu, Guangqing Chen, Jianjun Liu, Shuoye Zhou, Yanling Zhang, Aiguo Wang, Jianwei Wang, Ding Yan, Xianjie Cai, Junying Li, Xuchu Chen, Qiuying Li, Qiang Zeng, Weimin Guo, Yuanhui Wang, Regression prediction of tobacco chemical components during curing based on color quantification and machine learning, 2024, 14, 2045-2322, 10.1038/s41598-024-78426-y
    13. Xiaolong Yang, Hui Chang, Establishment and validation of a risk stratification model for stroke risk within three years in patients with cerebral small vessel disease using a combined MRI and machine learning algorithm, 2024, 29, 24726303, 100177, 10.1016/j.slast.2024.100177
    14. Carlos Matsinhe, Shingirai Brenda Kagodora, Tshifhiwa Mukheli, Tshepo Polly Mokoena, William Khabe Malebati, Maeyane Stephens Moeng, Thifhelimbilu Emmanuel Luvhengo, Machine Learning Algorithm-Aided Determination of Predictors of Mortality from Diabetic Foot Sepsis at a Regional Hospital in South Africa During the COVID-19 Pandemic, 2024, 60, 1648-9144, 1718, 10.3390/medicina60101718
    15. YongLong Chen, Weilong Wu, Houfu Dai, Overview of titanium alloy cutting based on machine learning, 2023, 126, 0268-3768, 4749, 10.1007/s00170-023-11475-1
    16. Rafał Tkaczyk, Grzegorz Madejski, Dawid Gradolewski, Damian Dziak, Wlodek J. Kulesza, Methodological Selection of Optimal Features for Object Classification Based on Stereovision System, 2024, 24, 1424-8220, 3941, 10.3390/s24123941
    17. Fang Zhang, 2024, English Deep Learning Model Based on Improved SM-2 Algorithm, 979-8-3503-8295-2, 1, 10.1109/ICDCOT61034.2024.10516205
    18. Rachna Agarwal, Shipra Agarwal, Varsha Mittal, 2024, Smart Commerce: Unleashing Machine Learning for Optimal Customer Experiences, 979-8-3503-7809-2, 1, 10.1109/ICEECT61758.2024.10739142
    19. Onome Ejeromedoghene, Moses Kumi, Ephraim Akor, Zexin Zhang, The application of machine learning in 3D/4D printed stimuli-responsive hydrogels, 2025, 336, 00018686, 103360, 10.1016/j.cis.2024.103360
    20. Aryan Anil Yamde, Vikesh Gurudas Lade, Ankush Babarao Bindwal, Manishkumar S. Tiwari, Ramesh Pandharinath Birmod, Machine learning approaches for the prediction of hydrogen uptake in metal-organic-frameworks: A comprehensive review, 2025, 98, 03603199, 1131, 10.1016/j.ijhydene.2024.12.131
    21. Lemya Sainudeen Attoor, Chitra R, A. M Anusha Bamini, 2024, A Comparative Analysis on Fashion Cloth Feature Extraction Methods using Text and Image Data, 979-8-3503-7913-6, 1, 10.1109/ICACC63692.2024.10845447
    22. Ruian Tang, Jianyu Yang, Changfa Shao, Ning Shen, Bo Chen, Yu Gu, Changming Li, Dong Xu, Chunxian Guo, Two-dimensional nanomaterials-based optical biosensors empowered by machine learning for intelligent diagnosis, 2025, 185, 01659936, 118162, 10.1016/j.trac.2025.118162
    23. Yao Jin, Yuan Ren, Xiang Xu, Zhao-Yuan Guo, Chong-Yuan Guo, Chong Li, A Deep Learning Estimation Method for Temperature-Induced Girder End Displacements of Suspension Bridges, 2025, 19, 1930-2991, 307, 10.32604/sdhm.2024.055265
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(3775) PDF downloads(468) Cited by(8)

Figures and Tables

Figures(1)  /  Tables(6)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog