MartialTerran commited on
Commit
582b308
·
verified ·
1 Parent(s): b93df3b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -2
README.md CHANGED
@@ -13,7 +13,15 @@ Large language models, especially those based on transformer architectures, have
13
 
14
  The Invention:
15
 
16
- This invention provides a method and apparatus for dynamically reducing logit computation by selectively deactivating tokens at the output layer. This deactivation effectively reduces the dimensions of the unembedding matrix and bias vector used in logit computation, leading to significant computational savings. This is accomplished during the forward pass of the model, allowing for context-dependent token deactivation.
 
 
 
 
 
 
 
 
17
 
18
  Detailed Description of the Method:
19
 
@@ -38,8 +46,15 @@ Probability Distribution and Token Selection: A probability distribution is gene
38
  Iteration: Steps 1 through 5 are repeated until the desired length of text is generated.
39
 
40
  Detailed Description of the Apparatus:
 
 
 
 
 
 
 
41
 
42
- The apparatus comprises:
43
 
44
  Processing Unit: A processing unit (e.g., CPU, GPU, TPU) configured to execute the steps of the method described above.
45
 
 
13
 
14
  The Invention:
15
 
16
+ This invention provides a technique to make LLMs faster and more efficient by intelligently reducing the number of logit calculations during inference. It achieves this by dynamically identifying relevant words based on context and employing various optimization strategies like pruning, hierarchical computation, approximation, or specialized hardware.
17
+ This invention provides a method and apparatus for dynamically reducing logit computation by selectively deactivating tokens at the output layer. This deactivation effectively reduces the dimensions of the unembedding matrix and bias vector used in logit computation, leading to significant computational savings. This is accomplished during the forward pass of the model, allowing for context-dependent token deactivation.
18
+ In one embodiment, the method performs Dynamic Pruning/Filtering, such as:
19
+
20
+ Identify irrelevant words: The system will analyze the context of the input text and dynamically identify words or select pregrouped sets of words from the vocabulary that are unlikely/unwanted to be the next word in the sequence.
21
+
22
+ Skip logit calculation: Logits are only calculated for a subset of the vocabulary – those deemed relevant/wanted – significantly reducing computation.
23
+
24
+ Using Contextual clues: Techniques like previous words, sentence structure, topic modeling, or even sentiment analysis could be used to determine relevance/selection.
25
 
26
  Detailed Description of the Method:
27
 
 
46
  Iteration: Steps 1 through 5 are repeated until the desired length of text is generated.
47
 
48
  Detailed Description of the Apparatus:
49
+ The "apparatus" mentioned may be implemented to include one or more of:
50
+
51
+ Software modules: Implemented as part of an LLM inference engine.
52
+
53
+ Hardware accelerators: Specialized processors (e.g., GPUs, TPUs) optimized for LLM computations.
54
+
55
+ Combined software-hardware systems: Leveraging both specialized hardware and optimized software algorithms.
56
 
57
+ An exemplary embodiment of the apparatus comprises:
58
 
59
  Processing Unit: A processing unit (e.g., CPU, GPU, TPU) configured to execute the steps of the method described above.
60