Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

4090 24G显存 单卡是否可以运行? #1

Open
yuwxyun275 opened this issue Mar 8, 2025 · 3 comments
Open

4090 24G显存 单卡是否可以运行? #1

yuwxyun275 opened this issue Mar 8, 2025 · 3 comments

Comments

@yuwxyun275
Copy link

作者您好,我对您的研究工作很感兴趣并且很感谢您将代码开源。我想询问单张24G显存的4090是否可以部署该模型?

@DragonHou77
Copy link
Collaborator

您好,很高兴你能认可我们的工作。该基准及框架是通过api调用大模型,因此对显卡性能并没有过高的要求,完全可以部署在4090上!

@yuwxyun275
Copy link
Author

yuwxyun275 commented Mar 10, 2025

我调用deepseek的api运行的时间很慢是什么原因?请问运行eval_LLM.py文件正常需要多少时间?

def call_with_messages_deepseek(QA, save_list, model="deepseek-chat"):
api_key = ""
client = OpenAI(api_key=api_key, base_url="https://api.deepseek.com/v1")
question = QA['question']
# print(question)
while True:
try:
response = client.chat.completions.create(
model=model,
messages=[{"role": "user", "content": question}],
temperature=0.7,
max_tokens=1024,
stream=False
)
QA.update({'response': response.choices[0].message.content, 'result': True})
save_list.append(QA)
break
except Exception as e:
print(f"DeepSeek 请求失败: {e}")
time.sleep(2)

File:hyperedge_count_cot_bag_test:0/5: 0%| | 2/3500 [01:20<39:51:26, 41.02s/it]

@DragonHou77
Copy link
Collaborator

DragonHou77 commented Mar 10, 2025 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants