Skip to content
This repository has been archived by the owner on Oct 11, 2024. It is now read-only.

Commit

Permalink
format
Browse files Browse the repository at this point in the history
  • Loading branch information
tlrmchlsmth committed Jul 16, 2024
1 parent adb6713 commit b733a84
Show file tree
Hide file tree
Showing 3 changed files with 7 additions and 5 deletions.
6 changes: 4 additions & 2 deletions vllm/attention/backends/placeholder_attn.py
Original file line number Diff line number Diff line change
@@ -1,8 +1,10 @@
from dataclasses import dataclass
from typing import (List, Optional, Tuple, Type)
from typing import List, Optional, Tuple, Type

import torch

from vllm.attention.backends.abstract import (AttentionBackend, AttentionImpl,
AttentionMetadata)
import torch

# Placeholder attention backend for models like Mamba that don't have attention.
# Mainly exists to sidestep get_attn_backend.
Expand Down
2 changes: 1 addition & 1 deletion vllm/engine/llm_engine.py
Original file line number Diff line number Diff line change
Expand Up @@ -261,7 +261,7 @@ def __init__(

if not self.model_config.embedding_mode:
# For all decoders including attention-free models like mamba,
# this must call _initialize_kv_caches, as this is where model
# this must call _initialize_kv_caches, as this is where model
# warmup and CUDA graphs creation happens.
self._initialize_kv_caches()

Expand Down
4 changes: 2 additions & 2 deletions vllm/worker/model_runner.py
Original file line number Diff line number Diff line change
Expand Up @@ -23,8 +23,8 @@
FLASHINFER_WORKSPACE_BUFFER_SIZE = 0

from vllm.attention import AttentionMetadata, get_attn_backend
from vllm.attention.backends.placeholder_attn import PlaceholderAttentionBackend

from vllm.attention.backends.placeholder_attn import (
PlaceholderAttentionBackend)
from vllm.config import (CacheConfig, DeviceConfig, LoadConfig, LoRAConfig,
ModelConfig, MultiModalConfig, ParallelConfig,
PromptAdapterConfig, SchedulerConfig)
Expand Down

0 comments on commit b733a84

Please sign in to comment.