Appearance
Cutter Cheat Sheet
Overview
Cutter is a modern, free and open-source reverse engineering platform powered by the Rizin framework, designed to provide an intuitive and powerful graphical user interface for binary analysis and reverse engineering tasks. Developed as a Qt-based frontend to Rizin (formerly Radare2), Cutter combines the powerful analysis capabilities of command-line reverse engineering tools with a user-friendly graphical interface that makes advanced binary analysis accessible to both beginners and experienced reverse engineers. The platform has gained significant adoption in the cybersecurity community as a viable alternative to expensive commercial reverse engineering tools, offering comparable functionality without licensing costs.
The core strength of Cutter lies in its comprehensive analysis engine that can handle multiple architectures including x86, x64, ARM, MIPS, PowerPC, and many others, making it suitable for analyzing binaries from various platforms including Windows, Linux, macOS, Android, and embedded systems. Cutter provides advanced features such as automatic function detection, control flow graph generation, cross-references analysis, string analysis, and cryptographic function identification. The platform's modular architecture allows for extensive customization through plugins and scripts, enabling users to extend its functionality for specialized analysis tasks.
Cutter's modern interface design focuses on workflow efficiency, featuring multiple synchronized views including disassembly, hexdump, graph view, and decompiler output that can be arranged and customized according to user preferences. The platform integrates seamlessly with debugging capabilities, supports various file formats including PE, ELF, Mach-O, and raw binaries, and provides collaborative features for team-based reverse engineering projects. Its active development community and extensive documentation make it an excellent choice for malware analysis, vulnerability research, software security assessment, and educational purposes in reverse engineering and binary analysis.
Installation
Windows Installation
Installing Cutter on Windows systems:
bash
# Download from GitHub releases
# Visit: https://github.com/rizinorg/cutter/releases
# Download Windows installer
# cutter-v2.3.4-Windows-x86_64.exe
# Run installer as administrator
cutter-v2.3.4-Windows-x86_64.exe
# Alternative: Portable version
# Download: cutter-v2.3.4-Windows-x86_64.zip
# Extract to desired location
# Run cutter.exe
# Install via Chocolatey
choco install cutter
# Install via Scoop
scoop bucket add extras
scoop install cutter
# Verify installation
cutter --version
Linux Installation
Installing Cutter on Linux distributions:
bash
# Ubuntu/Debian installation
sudo apt update
sudo apt install cutter
# Alternative: Download AppImage
wget https://github.com/rizinorg/cutter/releases/download/v2.3.4/Cutter-v2.3.4-Linux-x86_64.AppImage
chmod +x Cutter-v2.3.4-Linux-x86_64.AppImage
./Cutter-v2.3.4-Linux-x86_64.AppImage
# Arch Linux installation
sudo pacman -S cutter
# Fedora installation
sudo dnf install cutter
# Build from source
git clone https://github.com/rizinorg/cutter.git
cd cutter
git submodule update --init --recursive
# Install dependencies
sudo apt install qt5-default libqt5svg5-dev qttools5-dev qttools5-dev-tools
# Build
mkdir build && cd build
cmake ..
make -j$(nproc)
sudo make install
macOS Installation
bash
# Install via Homebrew
brew install --cask cutter
# Alternative: Download DMG
# Visit: https://github.com/rizinorg/cutter/releases
# Download: Cutter-v2.3.4-macOS-x86_64.dmg
# Install by dragging to Applications
# Build from source
git clone https://github.com/rizinorg/cutter.git
cd cutter
git submodule update --init --recursive
# Install dependencies
brew install qt5 cmake
# Build
mkdir build && cd build
cmake ..
make -j$(sysctl -n hw.ncpu)
Docker Installation
bash
# Create Cutter Docker environment
cat > Dockerfile << 'EOF'
FROM ubuntu:20.04
ENV DEBIAN_FRONTEND=noninteractive
RUN apt-get update && apt-get install -y \
cutter \
xvfb x11vnc fluxbox \
wget curl
# Setup VNC for GUI access
EXPOSE 5900
# Start script
COPY start.sh /start.sh
RUN chmod +x /start.sh
CMD ["/start.sh"]
EOF
# Create start script
cat > start.sh << 'EOF'
#!/bin/bash
Xvfb :1 -screen 0 1024x768x16 &
export DISPLAY=:1
fluxbox &
x11vnc -display :1 -nopw -listen localhost -xkb &
cutter
EOF
# Build and run
docker build -t cutter-re .
docker run -p 5900:5900 -v $(pwd)/samples:/samples cutter-re
Basic Usage
Opening and Loading Files
Basic file operations in Cutter:
bash
# Launch Cutter
cutter
# Open file via command line
cutter /path/to/binary
# Open file with specific options
cutter -A /path/to/binary # Auto-analysis
cutter -e bin.cache=true /path/to/binary # Enable caching
# Load file in Cutter GUI
# File -> Open File
# Select binary file
# Choose analysis options:
# - Auto-analysis level (0-4)
# - Architecture (if not auto-detected)
# - Bits (32/64)
# - Endianness
# - Base address
# Load raw binary
# File -> Open File
# Select "Raw binary" format
# Specify architecture and base address
# Load from URL
# File -> Open URL
# Enter URL to binary file
Basic Navigation
Navigating through the binary:
bash
# Navigation shortcuts
# G - Go to address/function
# Space - Switch between graph and linear view
# Tab - Switch between panels
# Ctrl+F - Search
# Ctrl+G - Go to address
# Address navigation
# Click on addresses in disassembly
# Use address bar at top
# Right-click -> "Go to" options
# Function navigation
# Functions panel (left sidebar)
# Click function name to navigate
# Use function list dropdown
# Cross-references
# Right-click on instruction
# "Show X-Refs" to see references
# "Show X-Refs to" to see what references this
# Bookmarks
# Right-click -> "Add bookmark"
# View -> Bookmarks panel
# Navigate to saved locations
Basic Analysis
Performing basic binary analysis:
bash
# Automatic analysis
# Analysis -> Auto Analysis
# Choose analysis level:
# - Level 0: Basic (fast)
# - Level 1: Advanced (recommended)
# - Level 2: Expert (slow but thorough)
# Manual analysis commands (in console)
# View -> Console to open Rizin console
# Basic information
i # File information
ii # Imports
ie # Exports
is # Symbols
iz # Strings
iS # Sections
# Function analysis
afl # List functions
af # Analyze function at current address
afi # Function information
afv # Function variables
# String analysis
izz # All strings
iz~password # Search for strings containing "password"
# Cross-references
axt # Cross-references to current address
axf # Cross-references from current address
Advanced Features
Graph View Analysis
Using the graph view for control flow analysis:
bash
# Switch to graph view
# Press Space or View -> Graph
# Graph navigation
# Mouse wheel - Zoom in/out
# Middle mouse drag - Pan
# Click nodes to navigate
# Double-click to enter function
# Graph layout options
# Right-click in graph area
# Layout options:
# - Hierarchical
# - Radial
# - Force-directed
# Minimap
# View -> Show Minimap
# Navigate large graphs quickly
# Graph analysis features
# Highlight paths between nodes
# Identify loops and branches
# Analyze function complexity
# Export graph as image
# Custom graph views
# Create custom graphs for specific analysis
# Filter nodes by criteria
# Focus on specific code paths
Decompiler Integration
Using the built-in decompiler:
bash
# Enable decompiler view
# View -> Decompiler
# Or press F5 in function
# Decompiler options
# Right-click in decompiler view
# Options:
# - Rename variables
# - Change variable types
# - Add comments
# - Export decompiled code
# Decompiler backends
# Preferences -> Decompiler
# Available backends:
# - Ghidra decompiler (r2ghidra)
# - RetDec
# - Snowman
# Synchronization
# Decompiler view syncs with disassembly
# Click in decompiler to highlight assembly
# Modifications reflect in both views
# Export decompiled code
# Right-click -> Export
# Save as C source file
# Include comments and annotations
Debugging Integration
Debugging capabilities in Cutter:
bash
# Start debugging session
# Debug -> Start Debug Session
# Choose debugger backend:
# - Native debugger
# - GDB
# - WinDbg (Windows)
# Set breakpoints
# Click on line number in disassembly
# Right-click -> Toggle breakpoint
# Conditional breakpoints available
# Debug controls
# F9 - Continue
# F10 - Step over
# F11 - Step into
# Shift+F11 - Step out
# Ctrl+F2 - Restart
# Watch variables
# Debug -> Registers panel
# Debug -> Stack panel
# Debug -> Memory panel
# Add custom watches
# Memory examination
# View -> Memory Map
# Examine memory regions
# Modify memory values
# Search memory patterns
Plugin System
Extending Cutter with plugins:
bash
# Plugin management
# Edit -> Preferences -> Plugins
# Enable/disable plugins
# Install new plugins
# Popular plugins
# r2ghidra - Ghidra decompiler integration
# r2dec - Alternative decompiler
# r2pipe - Python scripting
# r2yara - YARA rule integration
# Python scripting
# Tools -> Python Console
# Write custom analysis scripts
# Automate repetitive tasks
# Example Python script
import cutter
# Get current function
func = cutter.cmdj("afij")
print(f"Function: {func[0]['name']}")
# Get strings
strings = cutter.cmdj("izj")
for s in strings:
print(f"String: {s['string']}")
# Custom analysis
# Create custom analysis plugins
# Extend Cutter functionality
# Share plugins with community
Automation Scripts
Automated Malware Analysis
python
#!/usr/bin/env python3
# Automated malware analysis with Cutter
import cutter
import json
import os
import hashlib
from datetime import datetime
class CutterMalwareAnalyzer:
def __init__(self, sample_path):
self.sample_path = sample_path
self.analysis_results = {}
self.indicators = []
# Calculate file hash
with open(sample_path, 'rb') as f:
self.file_hash = hashlib.sha256(f.read()).hexdigest()
def basic_analysis(self):
"""Perform basic static analysis"""
print("Performing basic analysis...")
# File information
file_info = cutter.cmdj("ij")
self.analysis_results["file_info"] = file_info
# Sections
sections = cutter.cmdj("iSj")
self.analysis_results["sections"] = sections
# Imports
imports = cutter.cmdj("iij")
self.analysis_results["imports"] = imports
# Exports
exports = cutter.cmdj("iej")
self.analysis_results["exports"] = exports
# Strings
strings = cutter.cmdj("izzj")
self.analysis_results["strings"] = strings
print(f"Found {len(imports)} imports, {len(exports)} exports, {len(strings)} strings")
def function_analysis(self):
"""Analyze functions in the binary"""
print("Analyzing functions...")
# Auto-analyze functions
cutter.cmd("aaa")
# Get function list
functions = cutter.cmdj("aflj")
self.analysis_results["functions"] = functions
# Analyze suspicious functions
suspicious_functions = []
for func in functions:
func_name = func.get("name", "")
# Check for suspicious function names
suspicious_keywords = [
"crypt", "encode", "decode", "obfus", "pack",
"inject", "hook", "patch", "shell", "exec",
"download", "upload", "connect", "socket"
]
if any(keyword in func_name.lower() for keyword in suspicious_keywords):
suspicious_functions.append(func)
self.indicators.append({
"type": "suspicious_function",
"value": func_name,
"address": func.get("offset"),
"description": f"Suspicious function name: {func_name}"
})
self.analysis_results["suspicious_functions"] = suspicious_functions
print(f"Found {len(suspicious_functions)} suspicious functions")
def string_analysis(self):
"""Analyze strings for indicators"""
print("Analyzing strings...")
strings = self.analysis_results.get("strings", [])
# Suspicious string patterns
suspicious_patterns = [
r"http[s]?://", # URLs
r"\b\d{1,3}\.\d{1,3}\.\d{1,3}\.\d{1,3}\b", # IP addresses
r"[A-Za-z0-9+/]{20,}={0,2}", # Base64
r"\\x[0-9a-fA-F]{2}", # Hex encoded
r"cmd\.exe|powershell|bash|sh", # Shell commands
r"CreateProcess|ShellExecute|WinExec", # Process creation
r"RegOpenKey|RegSetValue|RegDeleteKey", # Registry operations
r"CreateFile|WriteFile|ReadFile", # File operations
r"socket|connect|send|recv", # Network operations
]
import re
suspicious_strings = []
for string_obj in strings:
string_value = string_obj.get("string", "")
for pattern in suspicious_patterns:
if re.search(pattern, string_value, re.IGNORECASE):
suspicious_strings.append(string_obj)
self.indicators.append({
"type": "suspicious_string",
"value": string_value,
"address": string_obj.get("vaddr"),
"pattern": pattern,
"description": f"Suspicious string matching pattern: {pattern}"
})
break
self.analysis_results["suspicious_strings"] = suspicious_strings
print(f"Found {len(suspicious_strings)} suspicious strings")
def import_analysis(self):
"""Analyze imports for suspicious APIs"""
print("Analyzing imports...")
imports = self.analysis_results.get("imports", [])
# Suspicious API categories
suspicious_apis = {
"process_injection": [
"CreateRemoteThread", "WriteProcessMemory", "VirtualAllocEx",
"OpenProcess", "NtCreateThreadEx", "RtlCreateUserThread"
],
"persistence": [
"RegSetValueEx", "RegCreateKeyEx", "CreateService",
"SetWindowsHookEx", "SetTimer"
],
"evasion": [
"IsDebuggerPresent", "CheckRemoteDebuggerPresent",
"GetTickCount", "QueryPerformanceCounter", "Sleep"
],
"network": [
"WSAStartup", "socket", "connect", "send", "recv",
"InternetOpen", "HttpOpenRequest", "HttpSendRequest"
],
"crypto": [
"CryptAcquireContext", "CryptCreateHash", "CryptEncrypt",
"CryptDecrypt", "CryptGenKey"
]
}
suspicious_imports = []
for import_obj in imports:
import_name = import_obj.get("name", "")
for category, apis in suspicious_apis.items():
if import_name in apis:
suspicious_imports.append({
"import": import_obj,
"category": category,
"api": import_name
})
self.indicators.append({
"type": "suspicious_import",
"value": import_name,
"category": category,
"description": f"Suspicious API import: {import_name} ({category})"
})
self.analysis_results["suspicious_imports"] = suspicious_imports
print(f"Found {len(suspicious_imports)} suspicious imports")
def entropy_analysis(self):
"""Analyze entropy of sections"""
print("Analyzing entropy...")
sections = self.analysis_results.get("sections", [])
high_entropy_sections = []
for section in sections:
# Get section data
section_name = section.get("name", "")
section_addr = section.get("vaddr", 0)
section_size = section.get("vsize", 0)
if section_size > 0:
# Calculate entropy (simplified)
try:
data = cutter.cmd(f"p8 {section_size} @ {section_addr}")
if data:
entropy = self.calculate_entropy(bytes.fromhex(data))
section["entropy"] = entropy
# High entropy might indicate packed/encrypted data
if entropy > 7.0:
high_entropy_sections.append(section)
self.indicators.append({
"type": "high_entropy_section",
"value": section_name,
"entropy": entropy,
"description": f"High entropy section: {section_name} (entropy: {entropy:.2f})"
})
except:
pass
self.analysis_results["high_entropy_sections"] = high_entropy_sections
print(f"Found {len(high_entropy_sections)} high entropy sections")
def calculate_entropy(self, data):
"""Calculate Shannon entropy of data"""
import math
from collections import Counter
if not data:
return 0
# Count byte frequencies
byte_counts = Counter(data)
data_len = len(data)
# Calculate entropy
entropy = 0
for count in byte_counts.values():
probability = count / data_len
entropy -= probability * math.log2(probability)
return entropy
def generate_report(self, output_file=None):
"""Generate analysis report"""
if not output_file:
output_file = f"malware_analysis_{self.file_hash[:8]}.json"
report = {
"analysis_info": {
"file_path": self.sample_path,
"file_hash": self.file_hash,
"timestamp": datetime.now().isoformat(),
"total_indicators": len(self.indicators)
},
"analysis_results": self.analysis_results,
"indicators": self.indicators,
"summary": {
"suspicious_functions": len(self.analysis_results.get("suspicious_functions", [])),
"suspicious_strings": len(self.analysis_results.get("suspicious_strings", [])),
"suspicious_imports": len(self.analysis_results.get("suspicious_imports", [])),
"high_entropy_sections": len(self.analysis_results.get("high_entropy_sections", []))
}
}
with open(output_file, 'w') as f:
json.dump(report, f, indent=2)
print(f"Analysis report saved: {output_file}")
return report
def run_full_analysis(self):
"""Run complete malware analysis"""
print(f"Starting malware analysis of: {self.sample_path}")
self.basic_analysis()
self.function_analysis()
self.string_analysis()
self.import_analysis()
self.entropy_analysis()
report = self.generate_report()
print(f"Analysis completed. Found {len(self.indicators)} indicators.")
return report
# Usage in Cutter
if __name__ == "__main__":
# This script should be run within Cutter's Python console
sample_path = "/path/to/malware/sample"
analyzer = CutterMalwareAnalyzer(sample_path)
report = analyzer.run_full_analysis()
Batch Binary Analysis
python
#!/usr/bin/env python3
# Batch binary analysis script
import os
import json
import subprocess
import hashlib
from datetime import datetime
from pathlib import Path
class CutterBatchAnalyzer:
def __init__(self, input_dir, output_dir):
self.input_dir = Path(input_dir)
self.output_dir = Path(output_dir)
self.output_dir.mkdir(exist_ok=True)
self.results = []
def analyze_binary(self, binary_path):
"""Analyze single binary with Cutter"""
print(f"Analyzing: {binary_path}")
# Calculate file hash
with open(binary_path, 'rb') as f:
file_hash = hashlib.sha256(f.read()).hexdigest()
# Create Cutter script for analysis
script_content = f"""
import cutter
import json
# Basic analysis
cutter.cmd("aaa")
# Collect information
results = {{
"file_info": cutter.cmdj("ij"),
"functions": cutter.cmdj("aflj"),
"imports": cutter.cmdj("iij"),
"exports": cutter.cmdj("iej"),
"strings": cutter.cmdj("izzj"),
"sections": cutter.cmdj("iSj")
}}
# Save results
with open("/tmp/cutter_results_{file_hash}.json", "w") as f:
json.dump(results, f, indent=2)
# Exit Cutter
cutter.cmd("q")
"""
script_path = f"/tmp/cutter_script_{file_hash}.py"
with open(script_path, 'w') as f:
f.write(script_content)
try:
# Run Cutter with script
cmd = [
"cutter",
"-A", # Auto-analysis
"-i", script_path, # Run script
str(binary_path)
]
result = subprocess.run(
cmd,
capture_output=True,
text=True,
timeout=300 # 5 minute timeout
)
# Load results
results_file = f"/tmp/cutter_results_{file_hash}.json"
if os.path.exists(results_file):
with open(results_file, 'r') as f:
analysis_results = json.load(f)
# Clean up temporary files
os.remove(script_path)
os.remove(results_file)
return {
"file_path": str(binary_path),
"file_hash": file_hash,
"status": "success",
"analysis_results": analysis_results,
"timestamp": datetime.now().isoformat()
}
else:
return {
"file_path": str(binary_path),
"file_hash": file_hash,
"status": "failed",
"error": "No results file generated",
"timestamp": datetime.now().isoformat()
}
except subprocess.TimeoutExpired:
return {
"file_path": str(binary_path),
"file_hash": file_hash,
"status": "timeout",
"error": "Analysis timed out",
"timestamp": datetime.now().isoformat()
}
except Exception as e:
return {
"file_path": str(binary_path),
"file_hash": file_hash,
"status": "error",
"error": str(e),
"timestamp": datetime.now().isoformat()
}
def find_binaries(self):
"""Find binary files in input directory"""
binary_extensions = ['.exe', '.dll', '.so', '.dylib', '.bin']
binaries = []
for file_path in self.input_dir.rglob('*'):
if file_path.is_file():
# Check by extension
if file_path.suffix.lower() in binary_extensions:
binaries.append(file_path)
# Check by file command
elif self.is_binary_file(file_path):
binaries.append(file_path)
return binaries
def is_binary_file(self, file_path):
"""Check if file is binary using file command"""
try:
result = subprocess.run(
['file', str(file_path)],
capture_output=True,
text=True
)
binary_indicators = [
'executable', 'ELF', 'PE32', 'Mach-O',
'shared object', 'dynamic library'
]
return any(indicator in result.stdout for indicator in binary_indicators)
except:
return False
def run_batch_analysis(self):
"""Run analysis on all binaries"""
binaries = self.find_binaries()
print(f"Found {len(binaries)} binary files to analyze")
for i, binary_path in enumerate(binaries, 1):
print(f"Progress: {i}/{len(binaries)}")
result = self.analyze_binary(binary_path)
self.results.append(result)
# Save individual result
result_file = self.output_dir / f"result_{result['file_hash'][:8]}.json"
with open(result_file, 'w') as f:
json.dump(result, f, indent=2)
# Generate summary report
self.generate_summary_report()
print(f"Batch analysis completed. Results saved in: {self.output_dir}")
def generate_summary_report(self):
"""Generate summary report"""
successful = len([r for r in self.results if r['status'] == 'success'])
failed = len([r for r in self.results if r['status'] == 'failed'])
timeout = len([r for r in self.results if r['status'] == 'timeout'])
error = len([r for r in self.results if r['status'] == 'error'])
summary = {
"batch_analysis_summary": {
"total_files": len(self.results),
"successful": successful,
"failed": failed,
"timeout": timeout,
"error": error,
"success_rate": (successful / len(self.results)) * 100 if self.results else 0
},
"results": self.results,
"timestamp": datetime.now().isoformat()
}
summary_file = self.output_dir / "batch_analysis_summary.json"
with open(summary_file, 'w') as f:
json.dump(summary, f, indent=2)
# Generate HTML report
self.generate_html_report(summary)
def generate_html_report(self, summary):
"""Generate HTML summary report"""
html_template = """
<!DOCTYPE html>
<html>
<head>
<title>Cutter Batch Analysis Report</title>
<style>
body { font-family: Arial, sans-serif; margin: 20px; }
.header { background-color: #f0f0f0; padding: 20px; }
.summary { background-color: #e6f3ff; padding: 15px; margin: 20px 0; }
.result { margin: 10px 0; padding: 10px; border-left: 4px solid #ccc; }
.success { border-left-color: #4caf50; }
.failed { border-left-color: #f44336; }
.timeout { border-left-color: #ff9800; }
.error { border-left-color: #9c27b0; }
</style>
</head>
<body>
<div class="header">
<h1>Cutter Batch Analysis Report</h1>
<p>Generated: {timestamp}</p>
</div>
<div class="summary">
<h2>Summary</h2>
<p>Total Files: {total_files}</p>
<p>Successful: {successful}</p>
<p>Failed: {failed}</p>
<p>Timeout: {timeout}</p>
<p>Error: {error}</p>
<p>Success Rate: {success_rate:.1f}%</p>
</div>
<h2>Results</h2>
{results_html}
</body>
</html>
"""
results_html = ""
for result in summary["results"]:
status_class = result["status"]
results_html += f"""
<div class="result {status_class}">
<h3>{os.path.basename(result['file_path'])}</h3>
<p>Status: {result['status'].upper()}</p>
<p>Hash: {result['file_hash']}</p>
<p>Timestamp: {result['timestamp']}</p>
{f"<p>Error: {result.get('error', '')}</p>" if 'error' in result else ""}
</div>
"""
html_content = html_template.format(
timestamp=summary["timestamp"],
total_files=summary["batch_analysis_summary"]["total_files"],
successful=summary["batch_analysis_summary"]["successful"],
failed=summary["batch_analysis_summary"]["failed"],
timeout=summary["batch_analysis_summary"]["timeout"],
error=summary["batch_analysis_summary"]["error"],
success_rate=summary["batch_analysis_summary"]["success_rate"],
results_html=results_html
)
html_file = self.output_dir / "batch_analysis_report.html"
with open(html_file, 'w') as f:
f.write(html_content)
# Usage
if __name__ == "__main__":
input_directory = "/path/to/binaries"
output_directory = "/path/to/results"
analyzer = CutterBatchAnalyzer(input_directory, output_directory)
analyzer.run_batch_analysis()
Function Signature Analysis
python
#!/usr/bin/env python3
# Function signature analysis and matching
import cutter
import json
import hashlib
class FunctionSignatureAnalyzer:
def __init__(self):
self.function_signatures = {}
self.known_signatures = self.load_known_signatures()
def load_known_signatures(self):
"""Load known function signatures database"""
# This would typically load from a database or file
# For demo purposes, we'll use a small set
return {
"crypto_functions": {
"md5_init": {
"pattern": "mov.*0x67452301",
"description": "MD5 initialization constant"
},
"sha1_init": {
"pattern": "mov.*0x67452301.*0xefcdab89",
"description": "SHA1 initialization constants"
},
"aes_sbox": {
"pattern": "0x63.*0x7c.*0x77.*0x7b",
"description": "AES S-box constants"
}
},
"compression": {
"zlib_header": {
"pattern": "0x78.*0x9c",
"description": "ZLIB header magic"
}
},
"network": {
"socket_init": {
"pattern": "WSAStartup.*0x0202",
"description": "Winsock initialization"
}
}
}
def extract_function_signature(self, func_addr):
"""Extract signature from function"""
# Get function information
func_info = cutter.cmdj(f"afij @ {func_addr}")
if not func_info:
return None
func_info = func_info[0]
func_size = func_info.get("size", 0)
if func_size == 0:
return None
# Get function bytes
func_bytes = cutter.cmd(f"p8 {func_size} @ {func_addr}")
# Calculate hash
func_hash = hashlib.md5(bytes.fromhex(func_bytes)).hexdigest()
# Get disassembly
disasm = cutter.cmd(f"pdf @ {func_addr}")
# Extract constants and patterns
constants = self.extract_constants(disasm)
patterns = self.extract_patterns(disasm)
signature = {
"address": func_addr,
"name": func_info.get("name", f"fcn.{func_addr:08x}"),
"size": func_size,
"hash": func_hash,
"constants": constants,
"patterns": patterns,
"disassembly": disasm
}
return signature
def extract_constants(self, disassembly):
"""Extract constants from disassembly"""
import re
constants = []
# Look for immediate values
const_patterns = [
r'0x[0-9a-fA-F]+', # Hex constants
r'\b\d+\b', # Decimal constants
]
for pattern in const_patterns:
matches = re.findall(pattern, disassembly)
constants.extend(matches)
# Remove duplicates and sort
return sorted(list(set(constants)))
def extract_patterns(self, disassembly):
"""Extract instruction patterns from disassembly"""
lines = disassembly.split('\n')
patterns = []
for line in lines:
# Extract instruction mnemonic
parts = line.strip().split()
if len(parts) >= 2:
instruction = parts[1] # Skip address
patterns.append(instruction)
return patterns
def match_signature(self, signature):
"""Match signature against known signatures"""
matches = []
for category, signatures in self.known_signatures.items():
for sig_name, sig_data in signatures.items():
pattern = sig_data["pattern"]
description = sig_data["description"]
# Check if pattern matches in disassembly
if pattern in signature["disassembly"]:
matches.append({
"category": category,
"name": sig_name,
"description": description,
"confidence": "high"
})
# Check constants
for const in signature["constants"]:
if const in pattern:
matches.append({
"category": category,
"name": sig_name,
"description": f"Constant match: {const}",
"confidence": "medium"
})
return matches
def analyze_all_functions(self):
"""Analyze all functions in the binary"""
print("Analyzing function signatures...")
# Get all functions
functions = cutter.cmdj("aflj")
results = []
for func in functions:
func_addr = func.get("offset")
func_name = func.get("name", "")
print(f"Analyzing function: {func_name} @ 0x{func_addr:x}")
# Extract signature
signature = self.extract_function_signature(func_addr)
if signature:
# Match against known signatures
matches = self.match_signature(signature)
result = {
"function": signature,
"matches": matches
}
results.append(result)
if matches:
print(f" Found {len(matches)} signature matches")
return results
def generate_signature_report(self, results, output_file="signature_analysis.json"):
"""Generate signature analysis report"""
# Count matches by category
category_counts = {}
total_matches = 0
for result in results:
for match in result["matches"]:
category = match["category"]
category_counts[category] = category_counts.get(category, 0) + 1
total_matches += 1
report = {
"signature_analysis": {
"total_functions": len(results),
"total_matches": total_matches,
"category_counts": category_counts
},
"results": results
}
with open(output_file, 'w') as f:
json.dump(report, f, indent=2)
print(f"Signature analysis report saved: {output_file}")
print(f"Total functions analyzed: {len(results)}")
print(f"Total signature matches: {total_matches}")
return report
# Usage in Cutter
if __name__ == "__main__":
analyzer = FunctionSignatureAnalyzer()
results = analyzer.analyze_all_functions()
report = analyzer.generate_signature_report(results)
Integration Examples
IDA Pro Migration
python
#!/usr/bin/env python3
# IDA Pro to Cutter migration helper
import cutter
import json
class IDACutterMigration:
def __init__(self):
self.ida_commands = {
# IDA command -> Cutter equivalent
"MakeCode": "af",
"MakeFunction": "af",
"MakeName": "afn",
"MakeComm": "CC",
"Jump": "s",
"GetFunctionName": "afi~name",
"GetString": "ps",
"FindBinary": "/x",
"GetBytes": "p8",
"PatchByte": "wx",
"ScreenEA": "s",
"here": "s",
"BADADDR": "0xffffffff"
}
def convert_ida_script(self, ida_script):
"""Convert IDA Python script to Cutter"""
# Basic conversion patterns
conversions = [
("idc.MakeCode", "cutter.cmd('af')"),
("idc.MakeFunction", "cutter.cmd('af')"),
("idc.GetFunctionName", "cutter.cmdj('afi')['name']"),
("idc.Jump", "cutter.cmd('s')"),
("idaapi.get_bytes", "cutter.cmd('p8')"),
("idc.here()", "cutter.cmd('s')"),
("print", "print") # Keep print statements
]
converted_script = ida_script
for ida_pattern, cutter_pattern in conversions:
converted_script = converted_script.replace(ida_pattern, cutter_pattern)
return converted_script
def export_ida_database(self, output_file="ida_export.json"):
"""Export IDA-like database information"""
# Collect information similar to IDA database
database = {
"functions": cutter.cmdj("aflj"),
"segments": cutter.cmdj("iSj"),
"imports": cutter.cmdj("iij"),
"exports": cutter.cmdj("iej"),
"strings": cutter.cmdj("izzj"),
"comments": self.get_all_comments(),
"names": self.get_all_names()
}
with open(output_file, 'w') as f:
json.dump(database, f, indent=2)
print(f"Database exported to: {output_file}")
return database
def get_all_comments(self):
"""Get all comments in the binary"""
# This would collect all comments
# Implementation depends on Cutter's comment system
return []
def get_all_names(self):
"""Get all named locations"""
# This would collect all named locations
# Implementation depends on Cutter's naming system
return []
# Usage
migration = IDACutterMigration()
database = migration.export_ida_database()
Ghidra Integration
python
#!/usr/bin/env python3
# Ghidra and Cutter integration
import cutter
import json
import subprocess
import tempfile
class GhidraCutterIntegration:
def __init__(self, ghidra_path="/opt/ghidra"):
self.ghidra_path = ghidra_path
def export_to_ghidra(self, binary_path, project_name):
"""Export binary to Ghidra project"""
# Create Ghidra headless script
script_content = f"""
import ghidra.app.util.importer.MessageLog;
import ghidra.app.util.Option;
import ghidra.app.util.bin.format.pe.PortableExecutable;
import ghidra.program.model.listing.Program;
import ghidra.util.task.TaskMonitor;
// Import binary
File binaryFile = new File("{binary_path}");
Program program = importProgram(binaryFile);
// Auto-analyze
analyzeProgram(program, TaskMonitor.DUMMY);
// Export analysis results
exportAnalysisResults(program, "{project_name}_analysis.json");
"""
# Run Ghidra headless
with tempfile.NamedTemporaryFile(mode='w', suffix='.java', delete=False) as f:
f.write(script_content)
script_path = f.name
cmd = [
f"{self.ghidra_path}/support/analyzeHeadless",
"/tmp/ghidra_projects",
project_name,
"-import", binary_path,
"-postScript", script_path
]
try:
result = subprocess.run(cmd, capture_output=True, text=True)
return result.returncode == 0
except Exception as e:
print(f"Error running Ghidra: {e}")
return False
def import_ghidra_analysis(self, analysis_file):
"""Import Ghidra analysis results into Cutter"""
try:
with open(analysis_file, 'r') as f:
ghidra_data = json.load(f)
# Import functions
if "functions" in ghidra_data:
for func in ghidra_data["functions"]:
addr = func.get("address")
name = func.get("name")
if addr and name:
cutter.cmd(f"af @ {addr}")
cutter.cmd(f"afn {name} @ {addr}")
# Import comments
if "comments" in ghidra_data:
for comment in ghidra_data["comments"]:
addr = comment.get("address")
text = comment.get("text")
if addr and text:
cutter.cmd(f"CC {text} @ {addr}")
print("Ghidra analysis imported successfully")
return True
except Exception as e:
print(f"Error importing Ghidra analysis: {e}")
return False
# Usage
ghidra_integration = GhidraCutterIntegration()
ghidra_integration.export_to_ghidra("/path/to/binary", "analysis_project")
Troubleshooting
Common Issues
Installation Problems:
bash
# Qt dependency issues
sudo apt install qt5-default libqt5svg5-dev
# Build dependency issues
sudo apt install cmake build-essential git
# Python plugin issues
pip install r2pipe
# AppImage execution issues
chmod +x Cutter-*.AppImage
./Cutter-*.AppImage --appimage-extract-and-run
Performance Issues:
bash
# Large binary analysis
# Disable auto-analysis for large files
cutter -A 0 large_binary.exe
# Memory usage optimization
# Limit analysis depth
# Use project files to save state
# Close unused views
# Graph rendering issues
# Reduce graph complexity
# Use linear view for large functions
# Adjust graph layout settings
Analysis Issues:
bash
# Function detection problems
# Manual function creation: af @ address
# Adjust analysis settings
# Use different analysis levels
# Decompiler issues
# Try different decompiler backends
# Check function boundaries
# Verify architecture detection
# Import/export problems
# Check file format support
# Verify file permissions
# Use appropriate import options
Debugging
Enable debugging and troubleshooting:
bash
# Verbose output
cutter -v binary_file
# Debug mode
cutter -d binary_file
# Console debugging
# View -> Console
# Use Rizin commands for debugging
# Log file analysis
# Check ~/.local/share/RadareOrg/Cutter/
# Review log files for errors
# Plugin debugging
# Check plugin loading in preferences
# Verify plugin compatibility
# Review plugin logs
Security Considerations
Safe Analysis Practices
Malware Analysis Safety:
- Use isolated virtual machines for malware analysis
- Disable network connectivity when analyzing malware
- Use snapshots to restore clean state
- Implement proper containment measures
- Monitor system behavior during analysis
Data Protection:
- Encrypt sensitive analysis results
- Secure storage of binary samples
- Implement access controls
- Regular backup of analysis data
- Secure disposal of temporary files
Legal and Ethical Considerations
Reverse Engineering Ethics:
- Respect software licenses and terms of service
- Comply with applicable laws and regulations
- Use reverse engineering for legitimate purposes
- Avoid copyright infringement
- Follow responsible disclosure practices
Best Practices:
- Document analysis methodology
- Maintain chain of custody for evidence
- Implement quality assurance processes
- Regular training and skill development
- Stay updated with legal requirements