Skip to content

Commit

Permalink
chore: remove repetitive words (#16957)
Browse files Browse the repository at this point in the history
  • Loading branch information
tianzedavid committed May 14, 2024
1 parent b7467aa commit cfe1711
Show file tree
Hide file tree
Showing 5 changed files with 7 additions and 7 deletions.
2 changes: 1 addition & 1 deletion gallery/how_to/deploy_models/deploy_prequantized.py
Original file line number Diff line number Diff line change
Expand Up @@ -162,7 +162,7 @@ def quantize_model(model, inp):
#
# You would see operators specific to quantization such as
# qnn.quantize, qnn.dequantize, qnn.requantize, and qnn.conv2d etc.
input_name = "input" # the input name can be be arbitrary for PyTorch frontend.
input_name = "input" # the input name can be arbitrary for PyTorch frontend.
input_shapes = [(input_name, (1, 3, 224, 224))]
mod, params = relay.frontend.from_pytorch(script_module, input_shapes)
# print(mod) # comment in to see the QNN IR dump
Expand Down
2 changes: 1 addition & 1 deletion include/tvm/relax/dataflow_pattern.h
Original file line number Diff line number Diff line change
Expand Up @@ -914,7 +914,7 @@ class ExternFuncPatternNode : public DFPatternNode {
public:
String global_symbol_; /*!< The global symbol name of the external function */

/*! \brief The the external function name */
/*! \brief The external function name */
const String& global_symbol() const { return global_symbol_; }
void VisitAttrs(tvm::AttrVisitor* v) { v->Visit("global_symbol", &global_symbol_); }

Expand Down
4 changes: 2 additions & 2 deletions src/runtime/contrib/vllm/attention_kernels.cu
Original file line number Diff line number Diff line change
Expand Up @@ -145,7 +145,7 @@ __device__ void paged_attention_kernel(

// Load the query to registers.
// Each thread in a thread group has a different part of the query.
// For example, if the the thread group size is 4, then the first thread in the group
// For example, if the thread group size is 4, then the first thread in the group
// has 0, 4, 8, ... th vectors of the query, and the second thread has 1, 5, 9, ...
// th vectors of the query, and so on.
// NOTE(woosuk): Because q is split from a qkv tensor, it may not be contiguous.
Expand Down Expand Up @@ -185,7 +185,7 @@ __device__ void paged_attention_kernel(

// Load a key to registers.
// Each thread in a thread group has a different part of the key.
// For example, if the the thread group size is 4, then the first thread in the group
// For example, if the thread group size is 4, then the first thread in the group
// has 0, 4, 8, ... th vectors of the key, and the second thread has 1, 5, 9, ... th
// vectors of the key, and so on.
for (int i = 0; i < NUM_TOKENS_PER_THREAD_GROUP; i++) {
Expand Down
2 changes: 1 addition & 1 deletion src/runtime/relax_vm/kv_state.h
Original file line number Diff line number Diff line change
Expand Up @@ -83,7 +83,7 @@ class KVStateObj : public Object {
* with prefill length "10", "15", "20", then we pass `[5, 1, 8]`
* as the seq_ids and `[10, 15, 20]` as the append_lengths.
* This method is invoked right before entering the model forward
* function, and contains operations to prepare the the incoming
* function, and contains operations to prepare the incoming
* forward. For instance, this method may send auxiliary KV cache
* data structures to GPUs so that they can be operated
* in the model forward function.
Expand Down
4 changes: 2 additions & 2 deletions src/runtime/relax_vm/paged_kv_cache.cc
Original file line number Diff line number Diff line change
Expand Up @@ -85,7 +85,7 @@ struct Block {
int32_t start_pos = 0;
/*!
* \brief The current attention sink length of the block.
* It means the the **first** sink size elements will be pinned
* It means the **first** sink size elements will be pinned
* in the KV cache even when sliding window is enabled.
*/
int32_t sink_length = 0;
Expand Down Expand Up @@ -247,7 +247,7 @@ class PagedKVCacheAuxDataManager {
/*!
* \brief Copy the append length indptr array on device.
* \note Since the Q/K/V data may have raggedness in terms of lengths,
* we represent the the append lengths in CSR format.
* we represent the append lengths in CSR format.
*/
virtual NDArray CopyCurAppendLengthIndptrAsync(std::vector<int32_t>* data) = 0;
/*! \brief Copy the k position offset of applying RoPE for each sequence. */
Expand Down

0 comments on commit cfe1711

Please sign in to comment.