Skip to content
Open
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
Issue fixes
  • Loading branch information
Tianyang-Zhang committed Mar 27, 2026
commit bf4ecc661761aee9943412c34128c33b463673e0
15 changes: 9 additions & 6 deletions .github/workflows/installation-test.yml
Original file line number Diff line number Diff line change
Expand Up @@ -73,8 +73,9 @@ jobs:
shell: bash
run: |
set -eo pipefail
whl_name=$(ls *server*.whl)
python -m pip install --find-links . "$whl_name"
common_whl=$(ls memmachine_common-*.whl | head -1)
server_whl=$(ls memmachine_server-*.whl | head -1)
python -m pip install --find-links . "$common_whl" "$server_whl"

#
# ─────────────────────────────────────
Expand Down Expand Up @@ -105,8 +106,9 @@ jobs:
run: |
set -eo pipefail
export PYTHONUTF8=1
whl_name=$(ls *server*.whl)
python -m pip install --find-links . "$whl_name"
common_whl=$(ls memmachine_common-*.whl | head -1)
server_whl=$(ls memmachine_server-*.whl | head -1)
python -m pip install --find-links . "$common_whl" "$server_whl"

#
# ─────────────────────────────────────
Expand All @@ -124,8 +126,9 @@ jobs:
shell: bash
run: |
set -eo pipefail
whl_name=$(ls *server*.whl)
python -m pip install --find-links . "$whl_name"
common_whl=$(ls memmachine_common-*.whl | head -1)
server_whl=$(ls memmachine_server-*.whl | head -1)
python -m pip install --find-links . "$common_whl" "$server_whl"

#
# ─────────────────────────────────────
Expand Down
5 changes: 3 additions & 2 deletions .github/workflows/test-server-package.yml
Original file line number Diff line number Diff line change
Expand Up @@ -52,8 +52,9 @@ jobs:
shell: bash
run: |
set -eo pipefail
whl_name=$(ls dist/memmachine_server-*.whl | head -1)
pip install --find-links dist/ "$whl_name"
common_whl=$(ls dist/memmachine_common-*.whl | head -1)
server_whl=$(ls dist/memmachine_server-*.whl | head -1)
pip install --find-links dist/ "$common_whl" "$server_whl"

- name: Test server package imports
shell: bash
Expand Down
4 changes: 1 addition & 3 deletions evaluation/retrieval_skill/benchmark_config.py
Original file line number Diff line number Diff line change
Expand Up @@ -16,14 +16,12 @@

from __future__ import annotations

import json
import os
import re
from dataclasses import dataclass, field
from dataclasses import dataclass
from pathlib import Path
from typing import Any


# ---------------------------------------------------------------------------
# Environment-variable substitution
# ---------------------------------------------------------------------------
Expand Down
12 changes: 10 additions & 2 deletions evaluation/retrieval_skill/llm_judge.py
Original file line number Diff line number Diff line change
Expand Up @@ -44,7 +44,12 @@
def _get_openai_client() -> openai.OpenAI:
global _openai_client
if _openai_client is None:
_openai_client = openai.OpenAI(api_key=os.getenv("OPENAI_API_KEY"))
api_key = os.getenv("OPENAI_API_KEY")
if not api_key:
raise RuntimeError(
"OPENAI_API_KEY must be set when no eval_llm client is provided."
)
_openai_client = openai.OpenAI(api_key=api_key)
return _openai_client


Expand Down Expand Up @@ -147,7 +152,10 @@ def main():
continue

label = evaluate_llm_judge(
question, gold_answer, generated_answer, args.model,
question,
gold_answer,
generated_answer,
args.model,
eval_llm=eval_llm,
)
LLM_JUDGE[category].append(label)
Expand Down
7 changes: 5 additions & 2 deletions evaluation/retrieval_skill/locomo_search.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,12 +7,13 @@
from typing import Any

from dotenv import load_dotenv
from memmachine_server.common.utils import async_with

REPO_ROOT = Path(__file__).resolve().parents[2]
if str(REPO_ROOT) not in sys.path:
sys.path.append(str(REPO_ROOT))

from memmachine_server.common.utils import async_with # noqa: E402

from evaluation.utils import skill_utils # noqa: E402

ANSWER_PROMPT = """
Expand Down Expand Up @@ -157,7 +158,9 @@ async def run_locomo( # noqa: C901
speaker = message["speaker"]
full_content.append(f"[{session_datetime}] {speaker}: {text}")

answer_model_name = answer_llm.model_name if answer_llm is not None else "gpt-5-mini"
answer_model_name = (
answer_llm.model_name if answer_llm is not None else "gpt-5-mini"
)
_, model, query_skill = await skill_utils.init_memmachine_params(
vector_graph_store=vector_graph_store,
session_id=group_id,
Expand Down
Loading
Loading