1. National Key Laboratory of Parallel and Distributed Computing, National University of Defense Technology,Changsha 410073, China
2. College of Computer, National University of Defense Technology, Changsha 410073, China
shenao@nudt.edu.cn
zqlai@nudt.edu.cn
dsli@nudt.edu.cn
Show less
History+
Received
Accepted
Published
2024-07-03
2024-10-30
Issue Date
Revised Date
2024-10-31
PDF
(2336KB)
Abstract
Low-precision training has emerged as a practical approach, saving the cost of time, memory, and energy during deep neural networks (DNNs) training. Typically, the use of lower precision introduces quantization errors that need to be minimized to maintain model performance, often neglecting to consider the potential benefits of reducing training precision. This paper rethinks low-precision training, highlighting the potential benefits of lowering precision: (1) low precision can serve as a form of regularization in DNN training by constraining excessive variance in the model; (2) layer-wise low precision can be seen as an alternative dimension of sparsity, orthogonal to pruning, contributing to improved generalization in DNNs. Based on these analyses, we propose a simple yet powerful technique – DPC (Decreasing Precision with layer Capacity), which directly assigns different bit-widths to model layers, without the need for an exhaustive analysis of the training process or any delicate low-precision criteria. Thorough extensive experiments on five datasets and fourteen models across various applications consistently demonstrate the effectiveness of the proposed DPC technique in saving computational cost (−16.21%–−44.37%) while achieving comparable or even superior accuracy (up to +0.68%, +0.21% on average). Furthermore, we offer feature embedding visualizations and conduct further analysis with experiments to investigate the underlying mechanisms behind DPC’s effectiveness, enhancing our understanding of low-precision training. Our source code will be released upon paper acceptance.
The remarkable performance achieved by modern deep neural networks (DNNs) relies on extensive training data and parameters, leading to substantial computational costs in training. Simultaneously, recent advancements in DNNs have spurred an increasing need for intelligent edge devices [1]. Many of these devices operate in dynamic real-world environments, where constrained resources render it impractical to bear the high training costs.
To address these challenges, low-precision training has emerged as a practical approach, employing low precision for both forward and backward propagation during DNN training to optimize the training process, thereby saving on time, memory, and energy costs [2,3]. Typically, the use of low precision leads to significant quantization errors, which should be minimized to maintain model performance. Thus, researchers have explored various solutions to mitigate the impact of quantization errors, such as adapted bit-width assignment [4–12], data format design [13–16], and quantizer optimization [3,17–23].
However, these approaches commonly regard low precision as a source of noise to be minimized or an performance-efficiency trade-off, without considering the potential benefits of low-precision training. Consequently, minimizing the low-precision noise during training requires complex analysis, which introduces additional computational overhead.
Recent findings reveal an opportunity to use low precision for simultaneously improving the efficiency and optimization of DNN training. For example, existing works show that noise can help DNN training theoretically or empirically, motivating us to rethink the role of low precision in DNN training [24,25]. Cyclic Precision Training (CPT) [26] adopts a dynamic precision strategy during training process, which simultaneously improves training efficiency and DNNs’ accuracy over state-of-the-art (SOTA) static low-precision training techniques. However, while CPT temporally schedules training precision, it does not investigate methods for assigning varying bit-widths in the spatial dimension of model layers. Meanwhile, random pruning with appropriate layer-wise sparsity ratios may achieve higher test accuracy [27–29]. However, pruning involves the sparsification of parameters, whereas low precision can be understood as the sparsification of data representation.
Motivated by these studies, our work seeks to address the question of how to leverage lower precision spatially to achieve the win-win in both training efficiency and model performance. We systematically investigate the potential benefits of lowering precision. Specifically, low-precision training can theoretically optimize the performance of DNNs and can serve as a regularization technique in DNN training, by constraining excessive variance within the model. We consider the layer-wise low precision as an alternative dimension of sparsity, orthogonal to pruning, which contributes to improved DNNs generalization.
Based on these analyses, we propose Decreasing Precision with layer Capacity (DPC), which determines the precision bound automatically and varies the precision according to layer capacity in a logarithmic schedule patterns. Surprisingly, we find that even without explicitly pursuing the reduction of errors associated with lower precision, low-precision training also has the opportunity to optimize model performance, thereby achieving a win-win for both DNN accuracy and training efficiency.
Thorough experiments on five datasets and fourteen models across various applications (including classification and language modeling) illustrate the consistent effectiveness of the proposed DPC technique in reducing training cost (−16.21%–−44.37%) while achieving comparable or even superior accuracy (up to +0.68%, +0.21% on average). Moreover, we offer feature embedding visualization and further analysis to explore the underlying mechanisms of DPC’s effectiveness, contributing to a better comprehension of low-precision training. Overall, our research and findings open up a new design knob for simultaneously improving the optimization and efficiency of DNN training.
Our contribution can be summarized as below:
1. We systematically investigated the potential benefits of low precision in training of DNNs: (i) theoretically demonstrating the enhancement of model generalization due to low precision; (ii) the regularization-like effects imparted by low precision; and (iii) the correlation between low precision and layer capacity.
2. We propose a simple yet powerful technique – DPC, which directly assigns different bit-widths to model layers based on layer capacity, without the need for an exhaustive analysis of the training process or any delicate low-precision criteria.
3. Experimental results validate the surprising effectiveness of DPC, demonstrating improvements in both efficiency (−16.21%–−44.37% training cost) and optimization performance (up to +0.68%, +0.21% on average). The consistency of these results across multiple trials underscores the method’s stability and reliability.
4. Further experimental analysis validates the potential benefits brought about by low precision during training, which aligns with our perspective and aids in a deeper understanding of low-precision training.
2 Related work
2.1 Sparsification in DNN
Quantization and pruning are two prominent strategies for sparsification in DNN training, aimed at enhancing the efficiency without significantly compromising accuracy. Quantization involves the use of low-bit precision to represent weights and activations, leading to a substantial reduction in model size and facilitating faster computation at the cost of some accuracy. To address this trade-off, recent studies have explored specialized quantizers and post-training quantization techniques to minimize inaccuracies [30–32]. An emerging direction is mixed-precision quantization, which applies different precision levels to various layers, striking a balance between computational efficiency and model accuracy [33–35]. Pruning, on the other hand, selectively removes redundant or less significant connections within DNNs to simplify the model. It shares similarities with low-precision training in that both aim to optimize DNN performance [27,36]. Additionally, many studies have provided optimization ideas for sparsification in DNN training. Research suggests that the volume of information can be reduced in a way that paradoxically improves model accuracy [37,38]. Understanding the dynamics of DNN training, including the role of noise and the sequential learning of features [24,25,39], provides further insights into how low-precision training might be optimized. These findings collectively invite exploration into new paradigms for more efficient and effective DNN training methodologies.
2.2 Low-precision training
2.2.1 Mitigating low-precision noise in training
Low-precision training, which employs low-bit quantization in both the forward and backward propagation phases, is aimed at enhancing the efficiency of the DNN training. However, the introduction of low precision often brings with it low-precision noise, which can affect the effectiveness of DNN training. Therefore, a key issue is how to reduce the impact of low-precision noise. Many studies [40–46] have analyzed the impact of low precision on DNN training and have proposed techniques to mitigate its negative effects, such as bit-width assignment [4–9,11,12,47], data format design [10,13–16], or quantizer optimization [3,17–23,48,49]. Moreover, to reduce excessive memory consumption during training, various compression techniques [50–52] have been employed to significantly reduce the size of activations. However, the attempts to minimize low-precision noise for different tasks and models often require complex analysis and computation, introducing additional computational overhead and lacking universality, which makes rapid task transfer difficult. At the same time, these methods focus primarily on minimizing errors caused by low precision to better trade-off training efficiency and performance. The potential benefits of low precision are overlooked.
2.2.2 Low precision for optimized DNN training
In contrast to minimizing the low-precision noise in DNN training, several studies have recently argued that low precision can actually enhance DNN training. Yang et al. [2] have treated precision as a hyperparameter in training. Zhuang et al. [53] have developed a method to train DNNs using both low and high precision to learn from each other. However, these methods require additional computational costs, which can impact the efficiency of these methods. To our knowledge, the most relevant work is CPT [26], which treats precision as analogous to the learning rate in DNN training and has demonstrated that employing variable bit widths during training can aid in locating flatter minima, thereby enhancing the model’s generalization. However, CPT focuses solely on temporal precision optimization and does not account for spatial precision considerations.
3 Methodology
In this section, we first introduce our rethinking of low-precision training that motivates us to develop DPC by analogy to regularization techniques and pruning in Section 3.1, and then present the DPC concept in Section 3.2 followed by the precision bounds exploration in Section 3.3, which aims to automate the precision schedule for DPC.
3.1 Rethinking of low-precision training
As mentioned in Section 2.2.1, the motivation for current research focuses on mitigating the impact of low precision in each training iteration. Paradoxically, some studies have achieved superior results with lower precision, despite aiming to align low-precision training with its high-precision counterpart [4,14,41]. In extreme cases, optimized results can even be attained through arbitrary use of low precision. Although some works mention the benefits of low precision in explaining their improvement, systematic research on the benefits is still lacking. The potential benefits of low precision may be underestimated. Considering that previous efforts to minimize the impact of low-precision noise often required complex analysis, we have attempted to focus directly on the potential benefits of low precision in DNN training, aiming to optimize the process without the need to deliberately minimize low-precision noise.
3.1.1 Low precision may improve the performance
This subsection provides the perspective from statistical learning to show that low precision may not hurt the performance of the neural networks, and sometimes even achieve improvement. Let us consider the problem for minimizing a model with small population risk :
where denotes the data, and is the data distribution, and is the loss function satisfying the constraint . For simplicity, is assumed to be convex with respect to . Eq. (1) is an ideal model, characterizing various tasks in the computer vision and artificial intelligence community. Because the distribution is usually unknown or very complicated, people use the empirical risk minimization (ERM) surrogate rather than directly solving (1). Given an i.i.d. sampling set , ERM is devoted to
The main workhorse for Eq. (2) is the stochastic gradient descent (SGD). In [Theorem 5.2, [54]], it has been proved that the generalization error of SGD is
where is the output of SGD after iterations. From Eq. (3), we can find that the upper bound is related to : a smaller indicates a better generalization ability. Recalling that is the upper bound of the norm of the gradients, that means smaller gradients yield better generalization. It is noteworthy that the gradient here is in the context of full precision training, directly reflecting the change in weights used in consecutive iterations. However, when in the context of low-precision training, the gradients are further approximated. Those tiny gradients which account for the majority are thus neglected by approximation. Consequently, when employing low-precision approximation, we use a gradient with smaller norms to some extent .
Thus, our theoretical analysis indicates that employing low precision not only augments the efficiency of DNN training but also presents the opportunity to attain DNN models with enhanced performance.
3.1.2 Low precision has a similar effect as regularization
DNNs used in practice are over-parametrized to the extent that they can easily fit random labels to the data, but models with increasing capacity will result in overfitting to the training data [55,56]. Regularization is a commonly used method to mitigate overfitting by reducing the variance of over-parameterized model [57], as explained in Fig.1(a). Similarly, low precision also restricts the model’s representation. Therefore, we hypothesize that low precision can be employed to reduce the variance, serving as a form of regularization technique in DNN training.
We verify this phenomenon through extensive experiments, and take the ResNet-38 on CIFAR-10 as an example in Fig.1(b). We show the variances of layers trained with three different precision configurations: The horizontal axis is the abbreviations for layers, e.g., G1B1C1 represents the 1st conv layer of the 1st block in the 1st group (the structural analysis of models is given in Appendix 1); The vertical axis represents the variance of layers (left, histogram), and the training bit-width of layers (right, dashed line). The blue configuration is the baseline where the entire model is trained using 8-bit, while the orange and green ones use lower precision (4-bit) at some different layers of the model. It can be seen that the variance of layers trained with lower precision is notably smaller than that of the baseline, indicating the regularization effect achieved by utilizing lower precision.
Therefore, our experimental observations reveal that low precision in DNN training exhibits similarities to regularization, thereby suggesting the potential benefits in enhancing test accuracy, even without deliberately minimizing quantization error.
3.1.3 Low precision for large capacity layers can help DNN generalization
The representational capacity of DNNs can be analyzed in two dimensions: the amount of data and the precision of each data representation [58]. DNN pruning in training, which can be seen as sparsification of the number of data, is a widely used technique to enhance DNN training [29,59]. Typically, the probability of pruning is proportional to the layer capacity [27,28,60]. Through pruning, DNNs can more effectively acquire sparse knowledge [27], breaking free from local minima [61], and thereby improving generalization [29]. Likewise, low precision can be considered as a means of sparsification in the dimension of the precision of each data. Thus, we hypothesize that low precision for large capacity layers can help DNN generalization.
We use WideResNet-38 as an example to analyze how to determine the relation between bit-width for layer and layer capacity to enhance test accuracy. Fig.2 shows four different training setups based on the layer capacity. The all 8-bit (blue) serves as our baseline, using 8-bit for the whole model. The latter three setups assign low precision to both activations and weights: fixed 6-bit (red) fixes the bit-width for layers at 6-bit. 8 to 4-bit (green) assigns lower bit-widths to higher capacity layers, while 4 to 8-bit (purple) does the opposite. Notably, as the first and last layers require very low computation, the computational costs for the latter three setups are similar.
It can be observed that only the final accuracy of 8 to 4-bit outperform the baseline, despite its lower accuracy compared to other configurations in the initial stage of training. Interestingly, pruning training also exhibits the power to augment final accuracy with a lower test accuracy in the initial stage [27]. These findings highlight the similarity between low precision and pruning, and setting low precision in line with pruning sparsity rates can enhance test accuracy.
To further verify the underlying mechanisms of this improvement, we compare our 8-bit baseline (all 8-bit) with decreases with layer capacity (8 to 4-bit). Fig.3 visualizes the feature embedding of last conv layer in ResNet-110. As expected, the clusters are mixed up together in the baseline model with fixed precision. While decreasing precision with layer capacity can split those clusters with more separable representations (especially the gap between automobile and deer & cat, gap between deer and frog & ship & cat), indicating a lower generalization error over the fixed high-precision baseline.
Thus, we discover that the bit-width selection for each layer in low-precision training shares similarities with the setting of pruning rates, suggesting that assigning higher sparsity (lower bit-width) to layers with larger capacity could potentially enhance the model’s generalization capability, thereby achieving higher accuracy.
3.2 Introduction of DPC
The key concept of DPC draws inspiration from our rethinking of low precision as well as recent findings [26] and [29]: The former explores the benefits of low precision temporally and advocates dynamic precision along the training trajectory. The latter demonstrates the strong effectiveness of sparse training in the dimension of parameter count along DNN layers, and identify that appropriate layer-wise sparsity ratios can be an important booster for training a randomly pruned network from scratch. Hence, we have reason to believe that we can leverage lower precision spatially to achieve the win-win in both DNN accuracy and efficiency. Here, we propose DPC, which allocates higher precision to the layers with smaller capacity and lower precision to the layers with larger capacity. Specifically, as shown in Fig.4, DPC varies the precision across layers within predefined thresholds, as opposed to imposing a static precision across the entire model. This facilitates the application of varying degrees of low-precision regularization to individual layers, thereby potentially enhancing the model’s overall performance.
While DPC can be implemented using different layer-wise scheduling methods, here we present as an example an implementation of DPC in a logarithmic manner:
where is the DPC precision for layer . and are the lower and upper precision bounds, respectively. denotes the rounding operation. is the number of parameters of layer , is the number of parameters of intermediate capacity layer, and are the number of parameters of the layer with the smallest capacity and the number of parameters of the layer with the largest capacity, respectively. is a hyperparameter controlling the interval and range of precision, as shown in Fig.4. Specifically, Eq. (4) outlines a way to allocate bit-widths for each layer. The bit-widths assigned by Eq. (4) fall between and . And uses a logarithmic function to implement the control of the bit-width based on the relative exponential/logarithmic relationship between the layer and the intermediate capacity layer. When is equal to , the bit-width is exactly equal to the intermediate value, i.e., . When is less than , then . And when is more than , then . further controls the range of bit-width . When the value of is smaller, the values of will be more concentrated, while when is larger, the upper and lower bounds of will be closer to and . We use precision bounds exploration to determine and (see Section 3.3).
In summary, after obtaining the model to be trained, our DPC method first employs precision bounds exploration (Section 3.3) to determine the upper and lower bounds of the training bit-width. Subsequently, based on the model layer capacities, it assigns training bit-widths to different layers according to the approach outlined in Eq. (4), resulting in the DPC model ready for optimized training.
In Section 4.3, we further determine , and find DPC is not sensitive to in this range. Additionally, we try to use other scheduling functions to make the precision decrease with layer capacity (i.e., not logarithmic schedule in Eq. (4)), and find that the logarithmic function has better performance and stability. We apply DPC to the weights and activations while maintaining a fixed precision for errors and gradients, the latter of which is to ensure the stability of the gradients [23]. More discussion can be found in Section 4.3.
3.3 Precision bounds exploration of DPC
In addition to the concept of DPC, the question of how to determine the range of the bit-width has not been addressed, i.e., and in Eq. (4). We propose precision bounds exploration to determine the upper and lower precision bounds of our DPC. Since there are no constraints on the upper precision bound, we use a viable training scheme as a fixed-precision counterpart, setting its bit-width as the upper precision bound, and DPC method further reduces the bit-width used in training. The precision bounds exploration employ a heuristic algorithm to quickly identify the lower precision bound through the application of cosine similarity at a negligible computational expense. The cosine similarity [10] describe the consistency between pre- and post-low-precision quantization (defined in Eq. (5), illustrated in Fig.5(a), where denotes the quantization function and represents the weight).
This metric serves as an indicator of the degradation of training performance, exhibiting a pronounced correlation with the ultimate accuracy [41]. For instance, in Fig.5(b), the accuracy of the final test remains almost unchanged as long as the cosine similarity is maintained at a high level. However, when the cosine similarity begins to decline significantly, the accuracy also tends to decrease correspondingly.
The exploration of precision bounds offers a simple method with negligible computational overhead for selecting the precision bound of DPC, as illustrated in Algorithm 1.
Fig.5(b) offers an intuitive explanation: specifically, we first train the fixed-precision counterpart for a number of iterations (fixed 8-bit, 640 iterations), then extract the weights of the largest capacity layer for analysis (in our DPC method, the largest capacity layer will be assigned the lowest bit-width). Subsequently, we quantize the weights of the layer to a lower bit-width, i.e., 7-bit, and calculate the cosine similarity, which is 0.9998. We then gradually reduce the bit-width and calculate the cosine similarity at each step until the value falls below the threshold (set at 0.95 in our experiments, with a cosine similarity of 0.9463 at 3-bit), indicating a noticeable error due to the low bit-width. We thus take the last bit-width above the threshold (i.e., 4-bit, with a cosine similarity of 0.9876 > 0.95 as shown in Fig.5(b)) as the lower bound of bit-width.
As mentioned above, we determined that the upper precision bound of DPC is the bit-width of the fixed-precision counterpart. We also find that DPC demonstrates efficacy across a spectrum of precision ranges, with varying upper bounds (Section 4.3).
4 Experiment
We present the experimental setup in Section 4.1, the overall performance to demonstrate the superiority of DPC in Section 4.2, then comprehensive ablation studies of DPC in Section 4.3, and further analysis to understand the benefits derived from DPC in Section 4.4.
4.1 Experiment setup
Models and datasets. We train ResNet [62] with varying depths and widths and MobileNet-V2 [63,64] on CIFAR-10/100 [65] and ImageNet [66]. For language modeling tasks, we follow [67] for Transformer on WikiText-103 and [68] for LSTM on PTB.
Training settings. Our experiments are based on CPT [26], and on this foundation, our DPC assigns training bit-widths for each layer, thereby training in the manner of DPC. We follow the standard training setting in all the experiments. Specifically, we follow SOTA settings in [69] for CIFAR-10/100 and [62] for ImageNet, respectively. Our experimental data are derived from four independent replications, with the average performance of each method recorded. An analysis of the variance and stability is presented in Section 4.3 and Appendix 2. We choose to employ the GMMLOWP quantization scheme, as detailed in Google’s open-source repository, which aligns with the methodologies utilized in CPT [26] and SBM [41]. The explanation of the quantizer is given in Appendix 3.
Precision configurations and baselines. We describe precision configurations in both spatial and temporal dimensions (We use CPT [26] to schedule the training precision in temporal dimension). Specifically, the precision of spatial represents the bit-width interval we calculated using the precision bounds exploration of DPC (method proposed in Section 3.3) and Eq. (4) of DPC. For example, spatial 8 to 4 / temporal 0 donates that the precision ranges from 8-bit to 4-bit as layer capacity increases, and during training, the precision of each layer does not change temporally. And spatial 8 / temporal 4 donates that the precision is fixed at 8-bit for each layer, while during training, the precision can be reduced by up to 4-bit circularly. In Section 4.2, temporal −50% donates the precision of each layer can be cycled down to half the precision of DPC temporally.
We apply our baselines on top of SBM [41], which are listed as follows:
● SBM: The whole DNN is trained with fixed precision, following the framework introduced in [41].
● Pruning: Random pruning techniques as described in [29], based on pre-determined layer-wise sparsities [27].
● CPT: The DNN is trained with dynamic precision, following the method introduced in [26].
● RPC: Precision is randomly assigned regardless of layer capacity.
● CPC: Precision remains constant as layer capacity increases.
● IPC: Precision increments as layer capacity increases.
● DPC: Our proposed method, where precision decreases as layer capacity increases.
● DPC+CPT: Precision for each layer is further cyclically reduced according to CPT, on top of our DPC.
4.2 Overall performance
Result on CIFAR.Tab.1 presents the experimental results on CIFAR-10/100, and shows that: (1) Our DPC achieves a win-win with both higher accuracy and lower training cost. (2) Compared with other spacial low-precision methods with similar training costs (i.e., RPC, CPC, IPC), DPC achieves higher accuracy consistently, indicating its appropriateness for regularizing DNN training. (3) DPC achieves higher accuracy compared to pruning. In contrast to pruning, low precision offers a gentler way to data representation rather than pruning it directly. (4) In most cases, DPC outperforms the temporal low-precision method CPT. Moreover, CPT requires the entire model to use high precision during inference, whereas DPC can use low precision layer-wise, reducing the memory requirements for hardware, which is critical for deep learning of edge devices. (5) We also test the method using lower precision in both spatial and temporal, DPC+CPT further improves the efficiency while surpassing the accuracy of the 8-bit baseline. This result suggests the potential of using low precision in both spatial and temporal dimensions to enhance DNN training. All the results presented above substantiate the efficacy of DPC, achieving a win-win by enhancing accuracy and reducing computational cost. These findings are also in accordance with our rethinking of the role of low precision in DNN training.
Feature embedding visualization: Fig.6 visualizes the feature embedding of last conv layer in wideResNet-38. We observe that with other methods, certain clusters tend to mix together with relatively small gap. In contrast, DPC splits clusters with wider spacing compared to other methods, indicating DPC helps DNN generalization with more separable feature representations. These results elucidate the reasons behind DPC’s ability to enhance accuracy: DPC optimizes the feature representation in DNN training, thereby improving the generalization of the model.
Result on ImageNet. We are also interested in how far we can go with DPC on ImageNet [66], a non-saturated dataset on which DNNs are less over-parameterized and are hard to be compressed [29]. As shown in Tab.2, all other methods trade off efficiency with accuracy, whereas DPC achieves comparable accuracy with higher efficiency, and sometimes even achieves higher accuracy. DPC+CPT further improves efficiency, but the excessively low precision leads to training underfitting, resulting in lower accuracy compared to the 8-bit baseline. Nevertheless, when compared to the SBM with similar training costs, DPC+CPT still exhibits higher accuracy, indicating the potential improvement in training with different precision in both spatial and temporal. The results indicate that, despite the increased complexity on the ImageNet, DPC is still capable of optimizing both accuracy and computational training costs.
DPC on full-precision basis: We also evaluate DPC’s performance when training based on full precision. As presented in Tab.3, DPC achieves a win-win in both efficiency and accuracy in most cases, even on ImageNet. DPC+CPT further improves efficiency by approximately half while maintaining comparable test accuracy (higher than other methods with similar efficiency). Compared with Tab.2, DPC achieves higher efficiency while maintaining comparable accuracy. The results demonstrate that DPC is equally effective when applied to a full-precision foundation. And the results is consistent with our insight that low-precision behaves like de-redundancy: Full precision has stronger data representation, there is more room for low-precision optimization.
Result on WikiText-103 and PTB.Tab.4 shows that (1) DPC achieves a win-win in both efficiency and model performance (perplexity, the lower the better), and (2) language modeling models/tasks are more sensitive to quantization, as it always adapts to a larger lower precision bound (consistent with observations [26,71]), due to the layers with similar capacities of natural language processing models. The results indicate that, while the performance gains may not be as pronounced in natural language processing models, DPC still manages to bring about optimizations in training.
DPC on VGG and AlexNet. Both ResNet and MobileNet-V2 use a well-designed structure in their blocks, which proved to have stronger representation capacity with similar amount of parameters. In Tab.5, we train AlexNet [72] and VGG [73] on CIFAR-10/100 to observe the effect of DPC on DNNs without special structural design. Tab.5 reveals the following insights: (1) Based on 8-bit, achieving higher accuracy with DPC becomes challenging. This suggests that relative to ResNet and MobileNet-V2, AlexNet and VGG-16 possess comparatively limited representation capabilities, potentially leading to some degree of underfitting with low precision. (2) Based on 16-bit, DPC outperforms the baseline. The increased precision at this level introduces a certain redundancy in representation capacity, enabling DPC to optimize effectively. (3) DPC remains effective in scenarios involving the combined use of convolutional and fully connected layers (in cases like AlexNet and VGG-16). The observations presented above validate the efficacy of DPC across various types of layers, concurrently aligning with the role of low precision in reducing redundancy.
4.3 Ablation studies of DPC
DPC with different precision ranges. We evaluate the performance of DPC across a wide range of upper precision bounds, which correspond to a different target efficiency. Fig.7 illustrates the test accuracy, training cost, and variance (size of circles) from four replicates. We can see that (1) regardless of the upper precision bounds, when compared to the fixed-precision counterpart, DPC consistently achieves a win-win with both higher accuracy and lower training cost, and (2) DPC even reduces accuracy variance, thereby aligning more closely with the practical objective of efficient training.
DPC with different precision intervals. In Eq. (4), different value of correspond to distinct precision intervals. We conduct experiments to evaluate the influence on training. As shown in the Tab.6, when is too small (i.e., ), the precision for layers becomes indistinguishable, leading to degradation of DPC performance. However, within a reasonable interval range (), optimized accuracy can be achieved. Therefore, we set .
DPC with different precision schedules. We evaluate DPC using other precision schedule patterns in addition to the logarithmic one, including a power schedule and a sine schedule. We exclude the linear scheduling based on capacity, as the capacities across layers often exhibit multiplicative relationships. Here, we show several possible scheduling patterns. We employ the ratio between layer capacities as a variable for scheduling and explore a triangular schedule:
and a power schedule:
In the power schedule, when is larger than 1, the precision for layers will change drastically, degrading the final accuracy. Therefore, we set to 0.5 for scheduling.
Tab.7 shows that logarithmic outperform other schedule patterns in terms of both accuracy and stability. Therefore, we chose the logarithmic schedule for DPC.
We suspect that the logarithmic function performs better because of the exponential-logarithmic relationship between bit-width and representation capabilities of data. We leave how to determine the optimal pattern for a given model and task as a future work.
DPC on top of gradients and errors. We opt not to implement DPC on gradients and errors, primarily due to two considerations. Firstly, the instability associated with low-precision gradients and errors, as reported in [23], may impede the model’s convergence. Secondly, the precision requirements for gradients and errors typically exceed those for weights and activations, which suggests that the efficiency gains from applying DPC to these components are marginal, as indicated in [26].
To substantiate this decision, we present the outcomes of experiments where DPC was applied to gradients and errors alongside weights and activations. As depicted in Tab.8, the application of DPC to gradients and errors offers negligible improvements in accuracy and efficiency when compared to a fixed precision baseline. Consequently, we have chosen to maintain a fixed precision for gradients and errors in all subsequent experimental evaluations.
4.4 Further analysis of DPC
DPC alleviates overfitting during training. When the model size reaches a certain threshold, the test accuracy does not improve with the increased model size due to overfitting. As shown in Tab.9, deploying a model with excessive parameters leads to a reduction in test accuracy. Conversely, DPC allows for scaling up the size of models while further improving accuracy without overfitting.
The primary factor contributing to low-precision optimization. Recent findings shows the differences in training between layers from different depths [74,75]. Therefore, it is important to determine whether the depth of the layer or the layer capacity represents the main factor influencing low-precision optimization. To answer this question, we introduced UniformResNet for analysis. This model adopts the same architecture as ResNet, with similar overall computations, but with equal capacity for each layer (details in Appendix 1). Here, we use CPD/IPD/DPD to represent Constant/Increase/Decrease Precision with Depth, respectively.
As shown in Tab.10, CPD achieves better results, and DPD is better than IPD. Thus, it seems that the depth of the layer and the layer capacity jointly determine the matching precision, and the latter is dominant. Meanwhile, the accuracy of UniformResNet is far inferior to that of original ResNet with similar computation, which shows the superiority of pyramid architecture of original network [62,63]. Under the reasonable structure, depth and capacity of layers are consistent from the precision perspective.
DPC alleviates the vanishing gradients. We observe that DPC alleviates the degradation caused by vanishing gradients. Typically, the initial layers are challenging to train due to small gradients [76]. In Fig.8, we present the weight gradients of layers from the initial shallow-depth group (e.g., B1C1: 1st convolution in 1st block) of ResNet-38 trained under different precision configurations. At epoch 100 (Fig.8(a)), gradient magnitudes of SBM, IPC and DPC exhibit minor discrepancies, with gradients still not considerably diminished. However, by epoch 150 (Fig.8(b)), the gradients have substantially diminished, notably showcasing a marked increase in the gradient of DPC in comparison to those of SBM and IPC.
This phenomenon highlights the fact that DPC enables shallow-depth layers to be adequately updated in the final stage, thus mitigating the issue of vanishing gradients.
The underlying cause of this phenomenon can be explained by the analysis in [19,52] (in Section 5).
5 Discussion
5.1 Scale of models
From all the experimental results, we observe that models with larger scale (deeper/wider) are more friendly to low precision (achieving higher accuracy with lower computational cost), which shares similarities with recent finding in pruning [29]. On one hand, large models have more parameters, thereby providing more opportunities to identify valuable features using low-precision techniques. Conversely, models with intensive structure necessitate a more fine-grained representation and more careful adjustment for each parameter, rendering them more sensitive to changes in precision. The conjecture is somewhat similar to lottery ticket hypothesis proposed in [60]. On the other hand, the use of low precision is also capable of suppressing interfering features, thereby facilitating the learning of crucial features.
5.2 Formulation of low-precision training
Recent studies [19,52] investigate how does low-precision quantization affect DNN training by analyzing gradients variance. In [Eq. (7) in [52]], the gradient variance for Quantization Aware Training (QAT) can be written as
The first term in is just the full precision gradient variance, and it accounts for the minibatch sampling. All the rest terms account for the noise of utilizing compressed context. Specifically, the term with is the variance introduced by utilizing the compressed context .
And the gradient variance for Fully Quantized Training (FQT) can be written as [Eq. (7) in [19]]:
where the first term is the variance of the QAT gradient. All the remaining variance terms come from gradient quantization, where each term is the variance imposed by the quantizer on layer to the gradient on layer . These formulations highlight that the distribution of each layer is not solely determined by its own; rather, the variance of each layer is also influenced by the quantization of other layers. During training, the entire model adjusts collectively to accommodate its low-precision parts.
From our analysis of the model structure in Appendix 1, it can be found that capacities of initial layers tend to be small. Consequently, under our DPC, higher precision is employed in these initial layers, while lower precision for deep layers. According to Eqs. (8) and (9), low precision introduces additional variance, and the variance of the initial layer compounds with the variance introduced by subsequent layers employing low precision. In the final training stages, gradients concentrate near zero [3,13]. Hence, augmenting variance under such circumstances serves to increase gradients, thus mitigating the issue of vanishing gradients.
6 Conclusion and future work
We systematically investigate the benefits of employing low-precision in DNN training, which can be proven to (i) optimize DNNs performance, serve as a form of (ii) regularization and (iii) layer-wise de-redundancy. Building upon these explorations, we introduce a simple yet powerful technique – DPC. Extensive experiments validate that comparable or even superior accuracy can be achieved even without deliberately minimizing quantization errors. Our technique and experimental results contribute to a deeper understanding of low-precision training, offering a new perspective for simultaneously optimizing efficiency and model performance.
Our future works will focus on identifying more theoretical grounds and further analyze low-precision methods across both spatial and temporal dimensions. Additionally, Large-scale Language Models (LLMs) demonstrate remarkable capabilities but often require substantial training costs. Current widely-used approaches focus on compressing the model Post-Training Quantization (PTQ) [77], with limited application of low-precision techniques during the pre-training phase. Therefore, exploring how to integrate the perspectives discussed in this paper with low-precision methods during the training of LLMs to enhance efficiency and potentially improve model performance is a promising research direction. Considering the substantial cost associated with pre-training LLMs, training LLMs directly with low precision is challenging. Therefore, it may be beneficial to initially attempt to optimize the fine-tuning of LLMs by combining PTQ with fine-tuning [78], and further distinguishing the bit-widths of different layers.
Du J, Shen M, Du Y. A distributed in-situ CNN inference system for IOT applications. In: Proceedings of the 38th IEEE International Conference on Computer Design (ICCD). 2020, 279–287
[2]
Yang C, Wu Z, Chee J, De Sa C, Udell M. How low can we go: trading memory for error in low-precision training. In: Proceedings of the 10th International Conference on Learning Representations. 2022
[3]
Zhu F, Gong R, Yu F, Liu X, Wang Y, Li Z, Yang X, Yan J. Towards unified int8 training for convolutional neural network. In: Proceedings of 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2020, 1966–1976
[4]
Micikevicius P, Narang S, Alben J, Diamos G, Elsen E, Garcia D, Ginsburg B, Houston M, Kuchaiev O, Venkatesh G, Wu H. Mixed precision training. In: Proceedings of the 6th International Conference on Learning Representations. 2018
[5]
Zhou S, Wu Y, Ni Z, Zhou X, Wen H, Zou Y. DoReFa-Net: training low bitwidth convolutional neural networks with low bitwidth gradients. 2016, arXiv preprint arXiv: 1606.06160
[6]
Yang L, Jin Q. FracBits: mixed precision quantization via fractional bit-widths. In: Proceedings of the 35th AAAI Conference on Artificial Intelligence. 2021, 10612–10620
[7]
E lthakeb A T, P illigundla P, M ireshghallah F, Y azdanbakhsh A, E smaeilzadeh H. ReLeQ: a reinforcement learning approach for automatic deep quantization of neural networks. IEEE Micro, 2020, 40( 5): 37–45
[8]
Yang H, Duan L, Chen Y, Li H. BSQ: exploring bit-level sparsity for mixed-precision neural network quantization. In: Proceedings of the 9th International Conference on Learning Representations, 2021
[9]
Ma Z, He J, Qiu J, Cao H, Wang Y, Sun Z, Zheng L, Wang H, Tang S, Zheng T, Lin J, Feng G, Huang Z, Gao J, Zeng A, Zhang J, Zhong R, Shi T, Liu S, Zheng W, Tang J, Yang H, Liu X, Zhai J, Chen W. BaGuaLu: targeting brain scale pretrained models with over 37 million cores. In: Proceedings of the 27th ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming. 2022, 192–204
[10]
Lee S, Park J, Jeon D. Toward efficient low-precision training: data format optimization and hysteresis quantization. In: Proceedings of the 10th International Conference on Learning Representations. 2021
[11]
Sun X, Wang N, Chen C Y, Ni J M, Agrawal A, Cui X, Venkataramani S, El Maghraoui K, Srinivasan V V, Gopalakrishnan K. Ultra-low precision 4-bit training of deep neural networks. In: Proceedings of the 34th International Conference on Neural Information Processing Systems. 2020, 33: 1796–1807
[12]
Fu Y, You H, Zhao Y, Wang Y, Li C, Gopalakrishnan K, Wang Z, Lin Y. FracTrain: fractionally squeezing bit savings both temporally and spatially for efficient DNN training. In: Proceedings of the 34th International Conference on Neural Information Processing Systems. 2020, 1017
[13]
Koster U, Webb T J, Wang X, Nassar M, Bansal A K, Constable W H, Elibol O H, Gray S, Hall S, Hornof L, Khosrowshahi A, Kloss C, Pai R J, Rao N. Flexpoint: an adaptive numerical format for efficient training of deep neural networks. In: Proceedings of the 31st International Conference on Neural Information Processing Systems. 2017, 1740–1750
[14]
Das D, Mellempudi N, Mudigere D, Kalamkar D, Avancha S, Banerjee K, Sridharan S, Vaidyanathan K, Kaul B, Georganas E, Heinecke A, Dubey P, Corbal J, Shustrov N, Dubtsov R, Fomenko E, Pirogov V. Mixed precision training of convolutional neural networks using integer operations. In: Proceedings of the 6th International Conference on Learning Representations. 2018
[15]
Fox S, Rasoulinezhad S, Faraone J, Leong P, Leong P. A block minifloat representation for training deep neural networks. In: Proceedings of the 9th International Conference on Learning Representations. 2020
[16]
Sun X, Choi J, Chen C Y, Wang N, Venkataramani S, Srinivasan V V, Cui X, Zhang W, Gopalakrishnan K. Hybrid 8-bit floating point (HFP8) training and inference for deep neural networks. In: Proceedings of the 33rd International Conference on Neural Information Processing Systems. 2019, 441
[17]
Zhang D, Yang J, Ye D, Hua G. LQ-Nets: learned quantization for highly accurate and compact deep neural networks. In: Proceedings of the 15th European Conference on Computer Vision-ECCV 2018. 2018, 373–390
[18]
Choi J, Wang Z, Venkataramani S, Chuang P I J, Srinivasan V, Gopalakrishnan K. PACT: parameterized clipping activation for quantized neural networks. In: Proceedings of the 6th International Conference on Learning Representations, 2018
[19]
Chen J, Gai Y, Yao Z, Mahoney M W, Gonzalez J E. A statistical framework for low-bitwidth training of deep neural networks. In: Proceedings of the 34th International Conference on Neural Information Processing Systems. 2020, 75
[20]
Fu F, Hu Y, He Y, Jiang J, Shao Y, Zhang C, Cui B. Don’t waste your bits! Squeeze activations and gradients for deep neural networks via TINYSCRIPT. In: Proceedings of the 37th International Conference on Machine Learning. 2020, 309
[21]
Wu S, Li G, Chen F, Shi L. Training and inference with integers in deep neural networks. In: Proceedings of the 6th International Conference on Learning Representations, 2018
[22]
Y ang Y, D eng L, W u S, Y an T, X ie Y, L i G. Training high-performance and large-scale deep neural networks with full 8-bit integers. Neural Networks, 2020, 125: 70–82
[23]
Wang N, Choi J, Brand D, Chen C Y, Gopalakrishnan K. Training deep neural networks with 8-bit floating point numbers. In: Proceedings of the 32nd International Conference on Neural Information Processing Systems. 2018, 7686–7695
[24]
G randvalet Y, C anu S, B oucheron S. Noise injection: theoretical prospects. Neural Computation, 1997, 9( 5): 1093–1108
[25]
Neelakantan A, Vilnis L, Le Q V, Sutskever I, Kaiser L, Kurach K, Martens J. Adding gradient noise improves learning for very deep networks. In: Proceedings of the 5th International Conference on Learning Representations. 2017
[26]
Fu Y, Guo H, Li M, Yang X, Ding Y, Chandra V, Lin Y. CPT: efficient deep neural network training via cyclic precision. In: Proceedings of the 9th International Conference on Learning Representations. 2021
[27]
M ocanu D C, M ocanu E, S tone P, N guyen P H, G ibescu M, L iotta A. Scalable training of artificial neural networks with adaptive sparse connectivity inspired by network science. Nature Communications, 2018, 9( 1): 2383
[28]
Evci U, Gale T, Menick J, Castro P S, Elsen E. Rigging the lottery: making all tickets winners. In: Proceedings of the 37th International Conference on Machine Learning. 2020, 2943–2952
[29]
Liu S, Chen T, Chen X, Shen L, Mocanu D C, Wang Z, Pechenizkiy M. The unreasonable effectiveness of random pruning: return of the most naive baseline for sparse training. In: Proceedings of the 10th International Conference on Learning Representations, ICLR 2022. 2022, 21
[30]
Zhou A, Yao A, Guo Y, Xu L, Chen Y. Incremental network quantization: towards lossless CNNs with low-precision weights. In: Proceedings of the 5th International Conference on Learning Representations, 2017
[31]
Krishnamoorthi R. Quantizing deep convolutional networks for efficient inference: a whitepaper. 2018, arXiv preprint arXiv: 1806.08342
[32]
Banner R, Nahshan Y, Soudry D. Post training 4-bit quantization of convolutional networks for rapid-deployment. In: Proceedings of the 33rd International Conference on Neural Information Processing Systems. 2019, 714
[33]
Wang K, Liu Z, Lin Y, Lin J, Han S. HAQ: hardware-aware automated quantization with mixed precision. In: Proceedings of 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2019, 8604–8612
[34]
Guo C, Tang J, Hu W, Leng J, Zhang C, Yang F, Liu Y, Guo M, Zhu Y. OliVe: accelerating large language models via hardware-friendly outlier-victim pair quantization. In: Proceedings of the 50th Annual International Symposium on Computer Architecture. 2023, 3
[35]
Guo C, Zhang C, Leng J, Liu Z, Yang F, Liu Y, Guo M, Zhu Y. ANT: exploiting adaptive numerical data type for low-bit deep neural network quantization. In: Proceedings of the 55th IEEE/ACM International Symposium on Microarchitecture (MICRO). 2022, 1414–1433
[36]
O nan A, K orukoğlu S, B ulut H. A hybrid ensemble pruning approach based on consensus clustering and multi-objective evolutionary algorithm for sentiment classification. Information Processing & Management, 2017, 53( 4): 814–833
[37]
Achille A, Rovere M, Soatto S. Critical learning periods in deep networks. In: Proceedings of the 7th International Conference on Learning Representations. 2018
[38]
A chille A, S oatto S. Emergence of invariance and disentanglement in deep representations. Journal of Machine Learning Research, 2018, 19( 1): 1947–1980
[39]
Raghu M, Poole B, Kleinberg J, Ganguli S, Sohl Dickstein J. On the expressive power of deep neural networks. In: Proceedings of the 34th International Conference on Machine Learning. 2017, 2847–2854
[40]
Martinez B, Yang J, Bulat A, Tzimiropoulos G. Training binary neural networks with real-to-binary convolutions. In: Proceedings of the 8th International Conference on Learning Representations. 2020
[41]
Banner R, Hubara I, Hoffer E, Soudry D. Scalable methods for 8-bit training of neural networks. In: Proceedings of the 32nd International Conference on Neural Information Processing systems. 2018, 5151–5159
[42]
Park E, Yoo S. PROFIT: a novel training method for sub-4-bit MobileNet models. In: Proceedings of the 16th European Conference on Computer Vision. 2020, 430–446
[43]
Yang G, Zhang T, Kirichenko P, Bai J, Wilson A G, De Sa C. SWALP: stochastic weight averaging in low precision training. In: Proceedings of the 36th International Conference on Machine Learning. 2019, 7015–7024
[44]
Han R, Si M, Demmel J, You Y. Dynamic scaling for low-precision learning. In: Proceedings of the 26th ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming. 2021, 480–482
[45]
Feng B, Wang Y, Geng T, Li A, Ding Y. APNN-TC: accelerating arbitrary precision neural networks on ampere GPU tensor cores. In: Proceedings of the SC21: International Conference for High Performance Computing, Networking, Storage and Analysis. 2021, 1–14
[46]
Knorr F, Thoman P, Fahringer T. ndzip-gpu: efficient lossless compression of scientific floating-point data on GPUs. In: SC21: Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis. 2021, 1–13
[47]
Savarese P, Yuan X, Li Y, Maire M. Not all bits have equal value: heterogeneous precisions via trainable noise. In: Proceedings of the 36th International Conference on Neural Information Processing Systems. 2022, 35769–35782
[48]
Xi H, Li C, Chen J, Zhu J. Training transformers with 4-bit integers. In: Proceedings of the 37th Conference on Neural Information Processing Systems (NeurIPS 2023). 2023
[49]
Huang X, Shen Z, Li S, Liu Z, Hu X, Wicaksana J, Xing E, Cheng K T. SDQ: stochastic differentiable quantization with mixed precision. In: Proceedings of the 39th International Conference on Machine Learning. 2022, 9295–9309
[50]
Chakrabarti A, Moseley B. Backprop with approximate activations for memory-efficient network training. In: Proceedings of the 33rd International Conference on Neural Information Processing Systems. 2019, 218
[51]
Jin S, Li G, Song S L, Tao D. A novel memory-efficient deep learning training framework via error-bounded lossy compression. In: Proceedings of the 26th ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming. 2021, 485–487
[52]
Chen J, Zheng L, Yao Z, Wang D, Stoica I, Mahoney M, Gonzalez J. ActNN: reducing training memory footprint via 2-bit activation compressed training. In: Proceedings of the 38th International Conference on Machine Learning. 2021, 1803–1813
[53]
Zhuang B, Shen C, Tan M, Liu L, Reid I. Towards effective low-bitwidth convolutional neural networks. In: Proceedings of 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2018, 7920–7928
[54]
Hardt M, Recht B, Singer Y. Train faster, generalize better: Stability of stochastic gradient descent. In: Proceedings of the 33rd International Conference on Machine Learning. 2016, 1225–1234
[55]
Z hang C, B engio S, H ardt M, R echt B, V inyals O. Understanding deep learning (still) requires rethinking generalization. Communications of the ACM, 2021, 64( 3): 107–115
[56]
Neyshabur B, Li Z, Bhojanapalli S, LeCun Y, Srebro N. Towards understanding the role of over-parametrization in generalization of neural networks. 2018, arXiv preprint arXiv: 180512076
[57]
Poggio T, Torre V, Koch C. Computational vision and regularization theory. In: Fischler M A, Firschein Q, eds. Readings in Computer Vision: Issues, Problem, Principles, and Paradigms. Los Altos: Morgan Kaufmann, 1987, 638–643
[58]
Dettmers T, Zettlemoyer L. The case for 4-bit precision: k-bit inference scaling laws. In: Proceedings of the 40th International Conference on Machine Learning. 2023, 370
[59]
Su J, Chen Y, Cai T, Wu T, Gao R, Wang L, Lee J D. Sanity-checking pruning methods: random tickets can win the jackpot. In: Proceedings of the 34th International Conference on Neural Information Processing Systems. 2020, 33: 1712
[60]
Frankle J, Carbin M. The lottery ticket hypothesis: finding sparse, trainable neural networks. In: Proceedings of the 7th International Conference on Learning Representations. 2019
[61]
Mostafa H, Wang X. Parameter efficient training of deep convolutional neural networks by dynamic sparse reparameterization. In: Proceedings of the 36th International Conference on Machine Learning. 2019, 4646–4655
[62]
He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition. 2016, 770–778
[63]
Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L C. MobileNetV2: inverted residuals and linear bottlenecks. In: Proceedings of 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2018, 4510–4520
[64]
Wang Y, Jiang Z, Chen X, Xu P, Zhao Y, Lin Y, Wang Z. E2-train: training state-of-the-art CNNs with over 80% energy savings. In: Proceedings of the 33rd International Conference on Neural Information Processing Systems. 2019, 462
[65]
Krizhevsky A. Learning multiple layers of features from tiny images. University of Toronto, 2012
[66]
Deng J, Dong W, Socher R, Li L J, Li K, Fei-Fei L. ImageNet: a large-scale hierarchical image database. In: Proceedings of 2009 IEEE Conference on Computer Vision and Pattern Recognition. 2009, 248–255
[67]
Baevski A, Auli M. Adaptive input representations for neural language modeling. In: Proceedings of the 7th International Conference on Learning Representations. 2019
[68]
Merity S, Keskar N S, Socher R. Regularizing and optimizing LSTM language models. In: Proceedings of the 6th International Conference on Learning Representations. 2018
[69]
Wang X, Yu F, Dou Z Y, Darrell T, Gonzalez J E. SkipNet: learning dynamic routing in convolutional networks. In: Proceedings of the 15th European Conference on Computer Vision-ECCV 2018. 2018, 420–436
[70]
Howard A G, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H. MobileNets: efficient convolutional neural networks for mobile vision applications. 2017, arXiv preprint arXiv: 1704.04861
[71]
Hou L, Zhu J, Kwok J, Gao F, Qin T, Liu T Y. Normalization helps training of quantized LSTM. In: Proceedings of the 33rd Neural Information Processing Systems. 2019, 660
[72]
K rizhevsky A, S utskever I, H inton G E. ImageNet classification with deep convolutional neural networks. Communications of the ACM, 2017, 60( 6): 84–90
[73]
Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. 2014, arXiv preprint arXiv: 1409.1556
[74]
He C, Li S, Soltanolkotabi M, Avestimehr S. PipeTransformer: automated elastic pipelining for distributed training of transformers. 2021, arXiv preprint arXiv: 2102.03161
[75]
Raghu M, Gilmer J, Yosinski J, Sohl-Dickstein J. SVCCA: singular vector canonical correlation analysis for deep learning dynamics and interpretability. In: Proceedings of the 31st International Conference on Neural Information Processing Systems. 2017, 6078–6087
[76]
Nielsen M A. Neural Networks and Deep Learning. San Francisco, CA, USA: Determination Press, 2015
[77]
Frantar E, Ashkboos S, Hoefler T, Alistarh D. GPTQ: accurate post-training quantization for generative pre-trained transformers. In: Proceedings of the 11th International Conference on Learning Representations. 2023
[78]
Dettmers T, Pagnoni A, Holtzman A, Zettlemoyer L. QLORA: efficient finetuning of quantized LLMs. In: Proceedings of the 37th International Conference on Neural Information Processing Systems. 2024, 441
[79]
Huang G, Liu Z, Van Der Maaten L, Weinberger K Q. Densely connected convolutional networks. In: Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition. 2017, 2261–2269
RIGHTS & PERMISSIONS
The Author(s) 2024. This article is published with open access at link.springer.com and journal.hep.com.cn
AI Summary 中Eng×
Note: Please be aware that the following content is generated by artificial intelligence. This website is not responsible for any consequences arising from the use of this content.