Skip to content

Conversation

@Xunzhuo
Copy link

@Xunzhuo Xunzhuo commented Mar 25, 2025

Second-Me CLI Interface Enhancement

The Second-Me CLI interface has been enhanced with emojis and improved formatting to provide a better user experience.

Command Structure

Each command now includes specific emojis and color-coded output:

Command Emoji Description
seme start 🚀 Start Second-Me services
seme stop 🛑 Stop all services
seme restart 🔄 Restart services
seme status 📊 Check services status
seme setup ⚙️ Set up environment

Improved Logging

Log messages have been enhanced with appropriate emojis:

  • ℹ️ [INFO] - Informational messages
  • ✅ [SUCCESS] - Success messages
  • ⚠️ [WARNING] - Warning messages
  • ❌ [ERROR] - Error messages

Section Headers

Section headers now use a consistent format with emojis:

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
  🔷 SECTION NAME
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

Status Command Example

The status command now provides a more visual representation of service status:

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
  🔷 📊 SERVICE STATUS
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

Backend Service:
  PID File: Running ✅ (PID: 12345, Process: python)
     ▶ python server.py

Frontend Service:
  PID File: Running ✅ (PID: 12346, Process: node)
     ▶ node --watch app.js

LLM Server Status:
  Port 8080: In use ✅ (PID: 12347, Process: llama-server)
     ▶ ./llama-server --model models/7B.gguf

Summary:
  🖥️  Backend: Running ✅
  🌐 Frontend: Running ✅
  🦙 LLM Server: Running ✅

Setup Command Example

The setup command provides clear progress indicators:

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
  🔷 ⚙️ SETTING UP PYTHON ENVIRONMENT
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

[2023-08-10 12:34:56] ℹ️ [INFO] Using conda environment: second-me
[2023-08-10 12:34:56] ℹ️ [INFO] Creating Conda environment: second-me
[2023-08-10 12:35:12] ✅ [SUCCESS] Conda environment second-me created successfully ✓
[2023-08-10 12:35:12] ℹ️ [INFO] Installing Python dependencies...
[2023-08-10 12:36:05] ✅ [SUCCESS] Python environment setup completed ✓

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
  🔷 🦙 BUILDING LLAMA.CPP
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

Additional Emojis

Component-specific emojis make it easier to identify what's being processed:

  • 🐍 Python environment
  • 🌐 Frontend services
  • 🦙 LLama.cpp and LLM services
  • ✓ Check mark for completed tasks

These enhancements create a more engaging and informative command-line experience for users of the Second-Me CLI tool.

@CLAassistant
Copy link

CLAassistant commented Mar 25, 2025

CLA assistant check
All committers have signed the CLA.

Signed-off-by: bitliu <bitliu@tencent.com>
@Xunzhuo Xunzhuo marked this pull request as draft March 25, 2025 09:04
@kevin-mindverse
Copy link
Contributor

Hey @Xunzhuo 👋

Thanks for your contribution! Unfortunately, I cannot merge this PR because it contains binary files. As a security measure, we need to be able to review all code changes, and binary files make this impossible.

Please remove the binary file from the PR so we can proceed with the review.

Really appreciate your understanding on this! 🙏

Let me know once you've updated the PR.

@ScarletttMoon
Copy link
Collaborator

Hi @Xunzhuo 👋,

Thank you so much for your contribution to this PR! Your work is really appreciated. If you haven’t already, feel free to join our Discord community here: Discord Invite Link — it's a great place to connect with our team and other contributors, share ideas, and stay up to date with the project! You can find me as @scarlettt_moon there!

Also, feel free to add me directly on WeChat (id: Neon_MoonS) if you’d like to chat more or have any questions!

Looking forward to connecting! 😊

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants