from llama_index.core import Settings, SimpleDirectoryReader, VectorStoreIndex
from llama_index.embeddings.huggingface import HuggingFaceEmbedding
from llama_index.llms.huggingface import HuggingFaceLLM
llm = HuggingFaceLLM(model_name="/Users/krian/PycharmProjects/llm-study/06-Model-Deploy/model/Qwen/Qwen3-0.6B",
tokenizer_name="/Users/krian/PycharmProjects/llm-study/06-Model-Deploy/model/Qwen/Qwen3-0.6B",
model_kwargs={
"trust_remote_code": True
}, tokenizer_kwargs={
"trust_remote_code": True
})
Settings.llm = llm
embedding = HuggingFaceEmbedding(
model_name="/Users/krian/PycharmProjects/llm-study/17-LlamaIndex-Demo/model/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2")
Settings.embed_model = embedding
documents = SimpleDirectoryReader("/Users/krian/PycharmProjects/llm-study/17-LlamaIndex-Demo/data").load_data()
index = VectorStoreIndex.from_documents(documents)
query_engine = index.as_query_engine()
response = query_engine.query("What's my name?")
print(f"LLM Response: {response}")