1
0
mirror of https://github.com/osmarks/osmarkscalculator.git synced 2024-10-30 03:26:16 +00:00

initial version

This commit is contained in:
osmarks 2022-11-10 17:31:44 +00:00
commit b8eddc0837
8 changed files with 1312 additions and 0 deletions

4
.gitignore vendored Normal file
View File

@ -0,0 +1,4 @@
/target
osmarkscalculator.zip
osmarkscalculator.tar
src.zip

167
Cargo.lock generated Normal file
View File

@ -0,0 +1,167 @@
# This file is automatically @generated by Cargo.
# It is not intended for manual editing.
version = 3
[[package]]
name = "anyhow"
version = "1.0.44"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "61604a8f862e1d5c3229fdd78f8b02c68dcf73a4c4b05fd636d12240aaa242c1"
[[package]]
name = "autocfg"
version = "1.0.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "cdb031dd78e28731d87d56cc8ffef4a8f36ca26c38fe2de700543e627f8a464a"
[[package]]
name = "cfg-if"
version = "1.0.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "baf1de4339761588bc0619e3cbc0120ee582ebb74b53b4efbf79117bd2da40fd"
[[package]]
name = "crossbeam-channel"
version = "0.5.2"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "e54ea8bc3fb1ee042f5aace6e3c6e025d3874866da222930f70ce62aceba0bfa"
dependencies = [
"cfg-if",
"crossbeam-utils",
]
[[package]]
name = "crossbeam-deque"
version = "0.8.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "6455c0ca19f0d2fbf751b908d5c55c1f5cbc65e03c4225427254b46890bdde1e"
dependencies = [
"cfg-if",
"crossbeam-epoch",
"crossbeam-utils",
]
[[package]]
name = "crossbeam-epoch"
version = "0.9.6"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "97242a70df9b89a65d0b6df3c4bf5b9ce03c5b7309019777fbde37e7537f8762"
dependencies = [
"cfg-if",
"crossbeam-utils",
"lazy_static",
"memoffset",
"scopeguard",
]
[[package]]
name = "crossbeam-utils"
version = "0.8.6"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "cfcae03edb34f947e64acdb1c33ec169824e20657e9ecb61cef6c8c74dcb8120"
dependencies = [
"cfg-if",
"lazy_static",
]
[[package]]
name = "either"
version = "1.6.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "e78d4f1cc4ae33bbfc157ed5d5a5ef3bc29227303d595861deb238fcec4e9457"
[[package]]
name = "hermit-abi"
version = "0.1.19"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "62b467343b94ba476dcb2500d242dadbb39557df889310ac77c5d99100aaac33"
dependencies = [
"libc",
]
[[package]]
name = "inlinable_string"
version = "0.1.14"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "3094308123a0e9fd59659ce45e22de9f53fc1d2ac6e1feb9fef988e4f76cad77"
[[package]]
name = "itertools"
version = "0.10.3"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "a9a9d19fa1e79b6215ff29b9d6880b706147f16e9b1dbb1e4e5947b5b02bc5e3"
dependencies = [
"either",
]
[[package]]
name = "lazy_static"
version = "1.4.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "e2abad23fbc42b3700f2f279844dc832adb2b2eb069b2df918f455c4e18cc646"
[[package]]
name = "libc"
version = "0.2.114"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "b0005d08a8f7b65fb8073cb697aa0b12b631ed251ce73d862ce50eeb52ce3b50"
[[package]]
name = "memoffset"
version = "0.6.5"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "5aa361d4faea93603064a027415f07bd8e1d5c88c9fbf68bf56a285428fd79ce"
dependencies = [
"autocfg",
]
[[package]]
name = "num_cpus"
version = "1.13.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "19e64526ebdee182341572e50e9ad03965aa510cd94427a4549448f285e957a1"
dependencies = [
"hermit-abi",
"libc",
]
[[package]]
name = "osmarkscalculator"
version = "0.1.0"
dependencies = [
"anyhow",
"inlinable_string",
"itertools",
"rayon",
]
[[package]]
name = "rayon"
version = "1.5.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "c06aca804d41dbc8ba42dfd964f0d01334eceb64314b9ecf7c5fad5188a06d90"
dependencies = [
"autocfg",
"crossbeam-deque",
"either",
"rayon-core",
]
[[package]]
name = "rayon-core"
version = "1.9.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "d78120e2c850279833f1dd3582f730c4ab53ed95aeaaaa862a2a5c71b1656d8e"
dependencies = [
"crossbeam-channel",
"crossbeam-deque",
"crossbeam-utils",
"lazy_static",
"num_cpus",
]
[[package]]
name = "scopeguard"
version = "1.1.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "d29ab0c6d3fc0ee92fe66e2d99f700eab17a8d57d1c1d3b748380fb20baa78cd"

12
Cargo.toml Normal file
View File

@ -0,0 +1,12 @@
[package]
name = "osmarkscalculator"
version = "0.1.0"
edition = "2018"
# See more keys and their definitions at https://doc.rust-lang.org/cargo/reference/manifest.html
[dependencies]
anyhow = "1"
inlinable_string = "0.1"
rayon = "1.5"
itertools = "0.10"

57
src/env.rs Normal file
View File

@ -0,0 +1,57 @@
use inlinable_string::InlinableString;
use std::collections::HashMap;
use std::sync::Arc;
use anyhow::Result;
use crate::value::Value;
#[derive(Debug, Clone)]
pub enum RuleResult {
Exp(Value),
Intrinsic(usize)
}
#[derive(Debug, Clone)]
pub struct Rule {
pub condition: Value,
pub result: RuleResult,
}
#[derive(Clone)]
pub struct Operation {
pub commutative: bool,
pub associative: bool,
}
pub type Bindings = HashMap<InlinableString, Value>;
pub type Ops = HashMap<InlinableString, Operation>;
pub type Ruleset = HashMap<InlinableString, Vec<Rule>>;
pub type Intrinsics = HashMap<usize, Box<dyn Fn(&Bindings) -> Result<Value> + Sync + Send>>;
#[derive(Clone)]
pub struct Env {
pub ops: Arc<Ops>,
pub ruleset: Vec<Arc<Ruleset>>,
pub intrinsics: Arc<Intrinsics>,
pub bindings: Bindings
}
impl Env {
// Get details about an operation, falling back to the default of not commutative/associative if none exists
pub fn get_op(&self, name: &InlinableString) -> Operation {
self.ops.get(name).map(Clone::clone).unwrap_or(Operation { commutative: false, associative: false })
}
// Make a new Env with this extra ruleset in it
pub fn with_ruleset(&self, ruleset: Arc<Ruleset>) -> Self {
let mut new_env = self.clone();
new_env.ruleset.push(ruleset);
new_env
}
pub fn with_bindings(&self, bindings: &Bindings) -> Self {
let mut new_env = self.clone();
new_env.bindings = bindings.clone();
new_env
}
}

633
src/main.rs Normal file
View File

@ -0,0 +1,633 @@
use anyhow::{Result, Context, bail};
use inlinable_string::InlinableString;
use std::collections::HashMap;
use std::io::BufRead;
use std::borrow::Cow;
use std::convert::TryInto;
use std::sync::Arc;
use rayon::prelude::*;
mod parse;
mod value;
mod util;
mod env;
use value::Value;
use env::{Rule, Ruleset, Env, Bindings, RuleResult, Operation};
// Main pattern matcher function;
fn match_and_bind(expr: &Value, rule: &Rule, env: &Env) -> Result<Option<Value>> {
fn go(expr: &Value, cond: &Value, env: &Env, already_bound: &Bindings) -> Result<Option<Bindings>> {
match (expr, cond) {
// numbers match themselves
(Value::Num(a), Value::Num(b)) => if a == b { Ok(Some(HashMap::new())) } else { Ok(None) },
// handle predicated value - check all predicates, succeed with binding if they match
(val, Value::Call(x, args)) if x == "#" => {
let preds = &args[1..];
let (mut success, mut bindings) = match go(val, &args[0], env, already_bound)? {
Some(bindings) => (true, bindings),
None => (false, already_bound.clone())
};
for pred in preds {
match pred {
// "Num" predicate matches successfully if something is a number
Value::Identifier(i) if i.as_ref() == "Num" => {
match val {
Value::Num(_) => (),
_ => success = false
}
},
// "Ident" does the same for idents
Value::Identifier(i) if i.as_ref() == "Ident" => {
match val {
Value::Identifier(_) => (),
_ => success = false
}
},
// Invert match success
Value::Identifier(i) if i.as_ref() == "Not" => {
success = !success
},
Value::Call(head, args) if head.as_ref() == "And" => {
// Try all patterns it's given, and if any fails then fail the match
for arg in args.iter() {
match go(val, arg, env, &bindings)? {
Some(new_bindings) => bindings.extend(new_bindings),
None => success = false
}
}
},
Value::Call(head, args) if head.as_ref() == "Eq" => {
// Evaluate all arguments and check if they are equal
let mut compare_against = None;
for arg in args.iter() {
let mut evaluated_value = arg.subst(&bindings);
run_rewrite(&mut evaluated_value, env).context("evaluating Eq predicate")?;
match compare_against {
Some(ref x) => if x != &evaluated_value {
success = false
},
None => compare_against = Some(evaluated_value)
}
}
},
Value::Call(head, args) if head.as_ref() == "Gte" => {
// Evaluate all arguments and do comparison.
let mut x = args[0].subst(&bindings);
let mut y = args[1].subst(&bindings);
run_rewrite(&mut x, env).context("evaluating Gte predicate")?;
run_rewrite(&mut y, env).context("evaluating Gte predicate")?;
success &= x >= y;
},
Value::Call(head, args) if head.as_ref() == "Or" => {
// Tries all patterns it's given and will set the match to successful if *any* of them works
for arg in args.iter() {
match go(val, arg, env, &bindings)? {
Some(new_bindings) => {
bindings.extend(new_bindings);
success = true
},
None => ()
}
}
},
_ => bail!("invalid predicate {:?}", pred)
}
}
Ok(match success {
true => Some(bindings),
false => None
})
},
(Value::Call(exp_head, exp_args), Value::Call(rule_head, rule_args)) => {
let mut exp_args = Cow::Borrowed(exp_args);
// Regardless of any special casing for associativity etc., different heads mean rules can never match
if exp_head != rule_head { return Ok(None) }
let op = env.get_op(exp_head);
// Copy bindings from the upper-level matching, so that things like "a+(b+a)" work.
let mut out_bindings = already_bound.clone();
// Special case for associative expressions: split off extra arguments into a new tree
if op.associative && rule_args.len() < exp_args.len() {
let exp_args = exp_args.to_mut();
let rest = exp_args.split_off(rule_args.len() - 1);
let rem = Value::Call(exp_head.clone(), rest);
exp_args.push(rem);
}
if rule_args.len() != exp_args.len() { return Ok(None) }
// Try and match all "adjacent" arguments to each other
for (rule_arg, exp_arg) in rule_args.iter().zip(&*exp_args) {
match go(exp_arg, rule_arg, env, &out_bindings)? {
Some(x) => out_bindings.extend(x),
None => return Ok(None)
}
}
Ok(Some(out_bindings))
},
// identifier pattern matches anything, unless the identifier has already been bound to something else
(x, Value::Identifier(a)) => {
if let Some(b) = already_bound.get(a) {
if b != x {
return Ok(None);
}
};
Ok(Some(vec![(a.clone(), x.clone())].into_iter().collect()))
},
// anything else doesn't match
_ => Ok(None)
}
}
// special case at top level of expression - try to match pattern to different subranges of input
match (expr, &rule.condition) {
// this only applies to matching one "call" against a "call" pattern (with same head)
(Value::Call(ehead, eargs), Value::Call(rhead, rargs)) => {
// and also only to associative operations
if env.get_op(ehead).associative && eargs.len() > rargs.len() && ehead == rhead {
// consider all possible subranges of the arguments of the appropriate length
for range_start in 0..=(eargs.len() - rargs.len()) {
// extract the arguments & convert into new Value
let c_args = eargs[range_start..range_start + rargs.len()].iter().cloned().collect();
let c_call = Value::Call(ehead.clone(), c_args);
// attempt to match the new subrange against the current rule
if let Some(r) = match_and_bind(&c_call, rule, env)? {
// generate new output with result
let mut new_args = Vec::with_capacity(3);
// add back extra start items
if range_start != 0 {
new_args.push(Value::Call(ehead.clone(), eargs[0..range_start].iter().cloned().collect()))
}
new_args.push(r);
// add back extra end items
if range_start + rargs.len() != eargs.len() {
new_args.push(Value::Call(ehead.clone(), eargs[range_start + rargs.len()..eargs.len()].iter().cloned().collect()))
}
let new_exp = Value::Call(ehead.clone(), new_args);
return Ok(Some(new_exp))
}
}
}
}
_ => ()
}
// substitute bindings from matching into either an intrinsic or the output of the rule
if let Some(bindings) = go(expr, &rule.condition, env, &HashMap::new())? {
Ok(Some(match &rule.result {
RuleResult::Intrinsic(id) => env.intrinsics.get(id).unwrap()(&bindings).with_context(|| format!("applying intrinsic {}", id))?,
RuleResult::Exp(e) => e.subst(&bindings)
}))
} else {
Ok(None)
}
}
// Sort any commutative expressions
fn canonical_sort(v: &mut Value, env: &Env) -> Result<()> {
match v {
Value::Call(head, args) => if env.get_op(head).commutative {
args.sort();
},
_ => ()
}
Ok(())
}
// Associative expression flattening.
fn flatten_tree(v: &mut Value, env: &Env) -> Result<()> {
match v {
Value::Call(head, args) => {
if env.get_op(head).associative {
// Not the most efficient algorithm, but does work.
// Repeatedly find the position of a flatten-able child node, and splice it into the argument list.
loop {
let mut move_pos = None;
for (i, child) in args.iter().enumerate() {
if let Some(child_head) = child.head() {
if *head == child_head {
move_pos = Some(i)
}
}
}
match move_pos {
Some(pos) => {
let removed = std::mem::replace(&mut args[pos], Value::Num(0));
// We know that removed will be a Call (because its head wasn't None earlier). Unfortunately, rustc does not know this.
match removed {
Value::Call(_, removed_child_args) => args.splice(pos..=pos, removed_child_args.into_iter()),
_ => unreachable!()
};
},
None => break
}
}
}
// Also do sorting after flattening, to avoid any weirdness with ordering.
canonical_sort(v, env)?;
return Ok(())
},
_ => return Ok(())
}
}
// Applies rewrite rulesets to an expression.
fn run_rewrite(v: &mut Value, env: &Env) -> Result<()> {
loop {
// Compare original and final hash instead of storing a copy of the original value and checking equality
// Collision probability is negligible and this is substantially faster than storing/comparing copies.
let original_hash = v.get_hash();
flatten_tree(v, env).context("flattening tree")?;
// Call expressions can be rewritten using pattern matching rules; identifiers can be substituted for bindings if available
match v {
Value::Call(head, args) => {
let head = head.clone();
// Rewrite sub-expressions using existing environment
args.par_iter_mut().try_for_each(|arg| run_rewrite(arg, env).with_context(|| format!("rewriting {}", arg.render_to_string(env))))?;
// Try to apply all applicable rules from all rulesets, in sequence
for ruleset in env.ruleset.iter() {
if let Some(rules) = ruleset.get(&head) {
// Within a ruleset, rules are applied backward. This is nicer for users using the program interactively.
for rule in rules.iter().rev() {
if let Some(result) = match_and_bind(v, rule, env).with_context(|| format!("applying rule {} -> {:?}", rule.condition.render_to_string(env), rule.result))? {
*v = result;
flatten_tree(v, env).context("flattening tree after rule application")?;
}
}
}
}
},
// Substitute in bindings which have been provided
Value::Identifier(ident) => {
match env.bindings.get(ident) {
Some(val) => {
*v = val.clone();
},
None => return Ok(())
}
},
_ => {
return Ok(())
}
}
if original_hash == v.get_hash() {
break
}
}
Ok(())
}
// Utility function for defining intrinsic functions for binary operators.
// Converts a function which does the actual operation to a function from bindings to a value.
fn wrap_binop<F: 'static + Fn(i128, i128) -> Result<i128> + Sync + Send>(op: F) -> Box<dyn Fn(&Bindings) -> Result<Value> + Sync + Send> {
Box::new(move |bindings: &Bindings| {
let a = bindings.get(&InlinableString::from("a")).context("binop missing first argument")?.assert_num("binop first argument")?;
let b = bindings.get(&InlinableString::from("b")).context("binop missing second argument")?.assert_num("binop second argument")?;
op(a, b).map(Value::Num)
})
}
// Provides a basic environment with operator commutativity/associativity operations and intrinsics.
fn make_initial_env() -> Env {
let mut ops = HashMap::new();
ops.insert(InlinableString::from("+"), Operation { commutative: true, associative: true });
ops.insert(InlinableString::from("*"), Operation { commutative: true, associative: true });
ops.insert(InlinableString::from("-"), Operation { commutative: false, associative: false });
ops.insert(InlinableString::from("/"), Operation { commutative: false, associative: false });
ops.insert(InlinableString::from("^"), Operation { commutative: false, associative: false });
ops.insert(InlinableString::from("="), Operation { commutative: false, associative: false });
ops.insert(InlinableString::from("#"), Operation { commutative: false, associative: true });
let ops = Arc::new(ops);
let mut intrinsics = HashMap::new();
intrinsics.insert(0, wrap_binop(|a, b| a.checked_add(b).context("integer overflow")));
intrinsics.insert(1, wrap_binop(|a, b| a.checked_sub(b).context("integer overflow")));
intrinsics.insert(2, wrap_binop(|a, b| a.checked_mul(b).context("integer overflow")));
intrinsics.insert(3, wrap_binop(|a, b| a.checked_div(b).context("division by zero")));
intrinsics.insert(4, wrap_binop(|a, b| {
// The "pow" function takes a usize (machine-sized unsigned integer) and an i128 may not fit into this, so an extra conversion is needed
Ok(a.pow(b.try_into()?))
}));
intrinsics.insert(5, Box::new(|bindings| {
// Substitute a single, given binding var=value into a target expression
let var = bindings.get(&InlinableString::from("var")).unwrap();
let value = bindings.get(&InlinableString::from("value")).unwrap();
let target = bindings.get(&InlinableString::from("target")).unwrap();
let name = var.assert_ident("Subst")?;
let mut new_bindings = HashMap::new();
new_bindings.insert(name, value.clone());
Ok(target.subst(&new_bindings))
}));
intrinsics.insert(6, wrap_binop(|a, b| a.checked_rem(b).context("division by zero")));
let intrinsics = Arc::new(intrinsics);
Env {
ruleset: vec![],
ops: ops.clone(),
intrinsics: intrinsics.clone(),
bindings: HashMap::new()
}
}
const BUILTINS: &str = "
SetStage[all]
a#Num + b#Num = Intrinsic[0]
a#Num - b#Num = Intrinsic[1]
a#Num * b#Num = Intrinsic[2]
a#Num / b#Num = Intrinsic[3]
a#Num ^ b#Num = Intrinsic[4]
Subst[var=value, target] = Intrinsic[5]
Mod[a#Num, b#Num] = Intrinsic[6]
PushRuleset[builtins]
";
const GENERAL_RULES: &str = "
SetStage[all]
(a*b#Num)+(a*c#Num) = (b+c)*a
Negate[a] = 0 - a
a^b*a^c = a^(b+c)
a^0 = 1
a^1 = a
(a^b)^c = a^(b*c)
0*a = 0
0+a = a
1*a = a
x/x = 1
(n*x)/x = n
PushRuleset[general_rules]
";
const NORMALIZATION_RULES: &str = "
SetStage[norm]
a/b = a*b^Negate[1]
a+b#Num*a = (b+1)*a
a^b#Num#Gte[b, 2] = a*a^(b-1)
a-c#Num*b = a+Negate[c]*b
a+a = 2*a
a*(b+c) = a*b+a*c
a-b = a+Negate[1]*b
PushRuleset[normalization]
";
const DENORMALIZATION_RULES: &str = "
SetStage[denorm]
a*a = a^2
a^b#Num*a = a^(b+1)
c+a*b#Num#Gte[0, b] = c-a*Negate[b]
PushRuleset[denormalization]
";
const DIFFERENTIATION_DEFINITION: &str = "
SetStage[all]
D[x, x] = 1
D[a#Num, x] = 0
D[f+g, x] = D[f, x] + D[g, x]
D[f*g, x] = D[f, x] * g + D[g, x] * f
D[a#Num*f, x] = a * D[f, x]
PushRuleset[differentiation]
";
const FACTOR_DEFINITION: &str = "
SetStage[post_norm]
Factor[x, a*x+b] = x * (a + Factor[x, b] / x)
PushRuleset[factor]
SetStage[pre_denorm]
Factor[x, a] = a
PushRuleset[factor_postprocess]
SetStage[denorm]
x^n/x = x^(n-1)
(a*x^n)/x = a*x^(n-1)
PushRuleset[factor_postpostprocess]
";
struct ImperativeCtx {
bindings: Bindings,
current_ruleset_stage: InlinableString,
current_ruleset: Ruleset,
rulesets: HashMap<InlinableString, Arc<Ruleset>>,
stages: Vec<(InlinableString, Vec<InlinableString>)>,
base_env: Env
}
impl ImperativeCtx {
// Make a new imperative context
// Stages are currently hardcoded, as adding a way to manage them would add lots of complexity
// for limited benefit
fn init() -> Self {
let stages = [
"pre_norm",
"norm",
"post_norm",
"pre_denorm",
"denorm",
"post_denorm"
].iter().map(|name| (InlinableString::from(*name), vec![])).collect();
ImperativeCtx {
bindings: HashMap::new(),
current_ruleset_stage: InlinableString::from("post_norm"),
current_ruleset: HashMap::new(),
rulesets: HashMap::new(),
stages,
base_env: make_initial_env()
}
}
// Insert a rule into the current ruleset; handles switching out the result for a relevant intrinsic use, generating possible reorderings, and inserting into the lookup map.
fn insert_rule(&mut self, condition: &Value, result_val: Value) -> Result<()> {
let result = match result_val {
Value::Call(head, args) if head == "Intrinsic" => RuleResult::Intrinsic(args[0].assert_num("Intrinsic ID")? as usize),
_ => RuleResult::Exp(result_val)
};
for rearrangement in condition.pattern_reorderings(&self.base_env).into_iter() {
let rule = Rule {
condition: rearrangement,
result: result.clone()
};
self.current_ruleset.entry(condition.head().unwrap()).or_insert_with(Vec::new).push(rule);
}
Ok(())
}
// Run a single statement (roughly, a line of user input) on the current context
fn eval_statement(&mut self, mut stmt: Value) -> Result<Option<Value>> {
match stmt {
// = sets a binding or generates a new rule.
Value::Call(head, args) if head.as_ref() == "=" => {
match &args[0] {
// Create a binding if the LHS (left hand side) is just an identifier
Value::Identifier(id) => {
let rhs = self.eval_statement(args[1].clone())?;
if let Some(val) = rhs.clone() {
self.bindings.insert(id.clone(), val);
}
Ok(rhs)
},
// If the LHS is a call, then a rule should be created instead.
Value::Call(_head, _args) => {
let rhs = self.eval_statement(args[1].clone())?;
if let Some(val) = rhs.clone() {
self.insert_rule(&args[0], val)?;
}
Ok(rhs)
},
// Rebinding numbers can only bring confusion, so it is not allowed.
// They also do not have a head, and so cannot be inserted into the ruleset anyway.
Value::Num(_) => bail!("You cannot rebind numbers")
}
},
// SetStage[] calls set the stage the current ruleset will be applied at
Value::Call(head, args) if head.as_ref() == "SetStage" => {
let stage = args[0].assert_ident("SetStage requires an identifier for stage")?;
if stage != "all" && None == self.stages.iter().position(|s| s.0 == stage) {
bail!("No such stage {}", stage);
}
self.current_ruleset_stage = stage;
Ok(None)
},
// Move the current ruleset from the "buffer" into the actual list of rules to be applied at each stage
Value::Call(head, args) if head.as_ref() == "PushRuleset" => {
let name = args[0].assert_ident("PushRuleset requires an identifier for ruleset name")?;
// Get ruleset and set the current one to empty
let ruleset = std::mem::replace(&mut self.current_ruleset, HashMap::new());
// Push ruleset to stages it specifies
for (stage_name, stage_rulesets) in self.stages.iter_mut() {
if *stage_name == self.current_ruleset_stage || self.current_ruleset_stage == "all" {
stage_rulesets.push(name.clone());
}
}
// Insert actual ruleset data under its name
self.rulesets.insert(name, Arc::new(ruleset));
Ok(None)
},
// Anything not special should just be repeatedly run through each rewrite stage.
_ => {
let env = self.base_env.with_bindings(&self.bindings);
for (stage_name, stage_rulesets) in self.stages.iter() {
// Add relevant rulesets to a new environment for this stage
let mut env = env.clone();
for ruleset in stage_rulesets.iter() {
env = env.with_ruleset(self.rulesets[ruleset].clone());
}
// Also add the current ruleset if applicable
if self.current_ruleset_stage == *stage_name || self.current_ruleset_stage == "all" {
env = env.with_ruleset(Arc::new(self.current_ruleset.clone()));
}
run_rewrite(&mut stmt, &env).with_context(|| format!("failed in {} stage", stage_name))?;
// If a ruleset is only meant to be applied in one particular stage, it shouldn't have any later stages applied to it,
// or the transformation it's meant to do may be undone
if self.current_ruleset_stage == *stage_name {
break
}
}
Ok(Some(stmt))
}
}
}
// Evaluate an entire "program" (multiple statements delineated by ; or newlines)
fn eval_program(&mut self, program: &str) -> Result<Option<Value>> {
let mut tokens = parse::lex(program)?;
let mut last_value = None;
loop {
// Split at the next break token
let remaining_tokens = tokens.iter().position(|x| *x == parse::Token::Break).map(|ix| tokens.split_off(ix + 1));
// Trim EOF/break tokens
match tokens[tokens.len() - 1] {
parse::Token::Break | parse::Token::EOF => tokens.truncate(tokens.len() - 1),
_ => ()
};
// If the statement/line isn't blank, readd EOF for the parser, parse into an AST then Value, and evaluate the statement
if tokens.len() > 0 {
tokens.push(parse::Token::EOF);
let value = Value::from_ast(parse::parse(tokens)?);
last_value = self.eval_statement(value)?;
}
// If there was no break after the current position, this is now done. Otherwise, move onto the new remaining tokens.
match remaining_tokens {
Some(t) => { tokens = t },
None => break
}
}
Ok(last_value)
}
}
fn main() -> Result<()> {
let mut ctx = ImperativeCtx::init();
ctx.eval_program(BUILTINS)?;
ctx.eval_program(GENERAL_RULES)?;
ctx.eval_program(FACTOR_DEFINITION)?;
ctx.eval_program(DENORMALIZATION_RULES)?;
ctx.eval_program(NORMALIZATION_RULES)?;
ctx.eval_program(DIFFERENTIATION_DEFINITION)?;
let stdin = std::io::stdin();
for line in stdin.lock().lines() {
let line = line?;
match ctx.eval_program(&line) {
Ok(Some(result)) => println!("{}", result.render_to_string(&ctx.base_env)),
Ok(None) => (),
Err(e) => println!("Error: {:?}", e)
}
}
Ok(())
}
#[cfg(test)]
mod test {
use crate::{ImperativeCtx, BUILTINS, GENERAL_RULES, NORMALIZATION_RULES, DENORMALIZATION_RULES, DIFFERENTIATION_DEFINITION, FACTOR_DEFINITION};
#[test]
fn end_to_end_tests() {
let mut ctx = ImperativeCtx::init();
ctx.eval_program(BUILTINS).unwrap();
ctx.eval_program(GENERAL_RULES).unwrap();
ctx.eval_program(FACTOR_DEFINITION).unwrap();
ctx.eval_program(DENORMALIZATION_RULES).unwrap();
ctx.eval_program(NORMALIZATION_RULES).unwrap();
ctx.eval_program(DIFFERENTIATION_DEFINITION).unwrap();
let test_cases = [
("Factor[x, x*3+x^2]", "(3+x)*x"),
("x^a/x^(a+1)", "x^Negate[1]"),
("Negate[a+b]", "Negate[1]*b-a"),
("Subst[x=4, x+4+4+4+4]", "20"),
("(a+b)*(c+d)*(e+f)", "a*c*e+a*c*f+a*d*e+a*d*f+b*c*e+b*c*f+b*d*e+b*d*f"),
("(12+55)^3-75+16/(2*2)+5+3*4", "300709"),
("D[3*x^3 + 6*x, x] ", "6+9*x^2"),
("Fib[n] = Fib[n-1] + Fib[n-2]
Fib[0] = 0
Fib[1] = 1
Fib[6]", "8"),
("Subst[b=a, b+a]", "2*a"),
("a = 7
b = Negate[4]
a + b", "3"),
("IsEven[x] = 0
IsEven[x#Eq[Mod[x, 2], 0]] = 1
IsEven[3] - IsEven[4]", "Negate[1]"),
("(a+b+c)^2", "2*a*b+2*a*c+2*b*c+a^2+b^2+c^2"),
("(x+2)^7", "128+2*x^6+12*x^5+12*x^6+16*x^3+16*x^5+24*x^4+24*x^5+32*x^2+32*x^3+32*x^5+128*x^2+256*x^4+448*x+512*x^2+512*x^3+x^7")
];
for (input, expected_result) in test_cases {
let lhs = ctx.eval_program(input).unwrap();
let lhs = lhs.as_ref().unwrap().render_to_string(&ctx.base_env);
println!("{} evaluated to {}; expected {}", input, lhs, expected_result);
assert_eq!(lhs, expected_result);
}
let error_cases = [
("1/0")
];
for error_case in error_cases {
if let Err(e) = ctx.eval_program(error_case) {
println!("{} produced error {:?}", error_case, e);
} else {
panic!("should have errored: {}", error_case)
}
}
println!("All tests passed.")
}
}

215
src/parse.rs Normal file
View File

@ -0,0 +1,215 @@
use std::str::FromStr;
use anyhow::{Result, anyhow, Context};
use std::fmt;
use inlinable_string::{InlinableString, StringExt};
use crate::util::char_to_string;
#[derive(Debug, PartialEq, Eq, Clone)]
pub enum Token { Number(InlinableString), OpenBracket, CloseBracket, Op(char), EOF, Identifier(InlinableString), OpenSqBracket, CloseSqBracket, Comma, Break }
#[derive(Debug)]
enum LexState {
Number(InlinableString),
Identifier(InlinableString),
None
}
// lexer
// converts input InlinableString to tokens
pub fn lex(input: &str) -> Result<Vec<Token>> {
let mut toks = vec![];
let mut state = LexState::None;
for (index, char) in input.chars().enumerate() {
state = match (char, state) {
// if digit seen, switch into number state (and commit existing one if relevant)
('0'..='9', LexState::None) => LexState::Number(char_to_string(char)),
('0'..='9', LexState::Identifier(s)) => {
toks.push(Token::Identifier(s));
LexState::Number(char_to_string(char))
},
('0'..='9', LexState::Number(mut n)) => {
n.push(char);
LexState::Number(n)
},
// if special character seen, commit existing state and push operator/bracket/comma token
('#' | '+' | '(' | ')' | '-' | '/' | '*' | '^' | '=' | '[' | ']' | ',', state) => {
match state {
LexState::Number(s) => toks.push(Token::Number(s)),
LexState::Identifier(s) => toks.push(Token::Identifier(s)),
_ => ()
};
toks.push(match char {
'(' => Token::OpenBracket, ')' => Token::CloseBracket,
'[' => Token::OpenSqBracket, ']' => Token::CloseSqBracket,
',' => Token::Comma,
a => Token::Op(a)
});
LexState::None
},
// semicolon or newline is break
(';' | '\n', state) => {
match state {
LexState::Number(s) => toks.push(Token::Number(s)),
LexState::Identifier(s) => toks.push(Token::Identifier(s)),
_ => ()
};
toks.push(Token::Break);
LexState::None
},
// ignore all whitespace
(' ', state) => state,
// treat all unknown characters as part of identifiers
(char, LexState::None) => { LexState::Identifier(char_to_string(char)) },
(char, LexState::Identifier(mut s)) => {
s.push(char);
LexState::Identifier(s)
}
(char, state) => return Err(anyhow!("got {} in state {:?} (char {})", char, state, index))
}
}
// commit last thing
match state {
LexState::Number(s) => toks.push(Token::Number(s)),
LexState::Identifier(s) => toks.push(Token::Identifier(s)),
_ => ()
};
toks.push(Token::EOF);
Ok(toks)
}
#[derive(Debug)]
pub enum ParseError {
Invalid(Token)
}
impl fmt::Display for ParseError {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
match self {
ParseError::Invalid(tok) => write!(f, "invalid token {:?}", tok)
}
}
}
impl std::error::Error for ParseError {}
// parser: contains the sequence of tokens being parsed and current position (index in that)
// no lookahead is supported/used
struct Parser {
tokens: Vec<Token>,
position: usize
}
// static table of precedence for each operator
pub fn precedence(c: char) -> usize {
match c {
'=' => 0,
'+' => 1,
'*' => 2,
'-' => 1,
'/' => 2,
'^' => 3,
'#' => 4,
c => panic!("invalid operator char {}", c)
}
}
fn is_left_associative(c: char) -> bool {
match c {
'^' => false,
_ => true
}
}
impl Parser {
// Parsing utility functions
// Get current token
fn current(&self) -> Token { self.tokens[self.position].clone() }
// Advance current token
fn advance(&mut self) { self.position += 1 }
// Match current token against predicate and propagate error if it is not matched
fn expect<T, F: Fn(&Token) -> Result<T, ()>>(&mut self, pred: F) -> Result<T, ParseError> {
let current = self.current();
match pred(&current) {
Ok(r) => {
self.advance();
Ok(r)
},
Err(()) => Err(ParseError::Invalid(current))
}
}
// Parse "leaf" expression: number, arbitrary expression in brackets, identifier, or call (fn[arg1, arg2])
fn parse_leaf(&mut self) -> Result<Ast, ParseError> {
let res = match self.current() {
Token::OpenBracket => {
self.advance();
let res = self.parse_expr(0)?;
self.expect(|t| if *t == Token::CloseBracket { Ok(()) } else { Err(()) })?;
res
},
Token::Number(n) => {
let x = i128::from_str(&n).unwrap();
self.advance();
Ast::Num(x)
},
Token::Identifier(s) => {
self.advance();
Ast::Identifier(s)
}
t => return Err(ParseError::Invalid(t))
};
// Detection of call syntax: if [ occurs, try and parse comma-separated arguments until a ]
if let Token::OpenSqBracket = self.current() {
self.advance();
let mut args = vec![];
loop {
args.push(self.parse_expr(0)?);
match self.expect(|t| if *t == Token::CloseSqBracket || *t == Token::Comma { Ok(t.clone()) } else { Err(()) })? {
Token::CloseSqBracket => break,
_ => ()
}
}
Ok(Ast::Call(Box::new(res), args))
} else {
Ok(res)
}
}
// Parse expression including operators, using precedence climbing
fn parse_expr(&mut self, prec: usize) -> Result<Ast, ParseError> {
let mut result = self.parse_leaf()?;
loop {
match self.current() {
Token::Op(op) if precedence(op) >= prec => {
self.advance();
let nested_precedence = if is_left_associative(op) { precedence(op) + 1 } else { precedence(op) };
let next = self.parse_expr(nested_precedence)?;
result = Ast::Op(op, Box::new(result), Box::new(next))
},
// Break out of loop if operator's precedence is lower, or nonoperator thing encountered
// This means that a lower-precedence operator will become part of the tree enclosing the current one
_ => break
};
};
Ok(result)
}
}
pub fn parse(t: Vec<Token>) -> Result<Ast> {
let mut parser = Parser {
tokens: t,
position: 0
};
// Provide slightly more helpful error message indicating the token which isn't valid
let result = parser.parse_expr(0).with_context(|| format!("at token {}", parser.position))?;
if parser.current() != Token::EOF { return Err(anyhow!("Expected EOF at end of token sequence")) }
Ok(result)
}
#[derive(Debug)]
pub enum Ast {
Num(i128),
Identifier(InlinableString),
Op(char, Box<Ast>, Box<Ast>),
Call(Box<Ast>, Vec<Ast>)
}

7
src/util.rs Normal file
View File

@ -0,0 +1,7 @@
use inlinable_string::{InlinableString, StringExt};
pub fn char_to_string(c: char) -> InlinableString {
let mut s = InlinableString::new();
s.push(c);
s
}

217
src/value.rs Normal file
View File

@ -0,0 +1,217 @@
use inlinable_string::{InlinableString, StringExt};
use std::hash::{Hash, Hasher};
use std::collections::{hash_map::DefaultHasher, HashSet, HashMap};
use std::fmt::{self, Write};
use itertools::Itertools;
use anyhow::{Result, anyhow};
use crate::parse::{Ast, precedence};
use crate::env::{Env, Bindings};
use crate::util::char_to_string;
#[derive(Debug, PartialEq, Eq, PartialOrd, Ord, Clone, Hash)]
pub enum Value {
Num(i128),
Call(InlinableString, Vec<Value>),
Identifier(InlinableString),
}
impl Value {
// Converts an AST from `parse.rs` to a Value
pub fn from_ast(ast: Ast) -> Self {
match ast {
Ast::Op(char, t1, t2) => Value::Call(char_to_string(char), vec![Value::from_ast(*t1), Value::from_ast(*t2)]),
Ast::Call(f, args) => {
Value::Call(match *f {
Ast::Identifier(n) => n,
_ => unimplemented!()
}, args.into_iter().map(Value::from_ast).collect())
},
Ast::Num(n) => Value::Num(n),
Ast::Identifier(i) => Value::Identifier(i)
}
}
// Gets the hash of a Value
pub fn get_hash(&self) -> u64 {
// according to https://doc.rust-lang.org/std/collections/hash_map/struct.DefaultHasher.html, all instances created here are guaranteed to be the same
let mut hasher = DefaultHasher::new();
self.hash(&mut hasher);
hasher.finish()
}
// Gets the head (string at start of call expression)
pub fn head(&self) -> Option<InlinableString> {
match self {
Value::Call(fun, _) => Some(fun.clone()),
_ => None
}
}
// Replace variables with incremental IDs, vaguely like de Bruijn indices in lambda calculus.
// Allows patterns to be compared regardless of the names of the identifiers they contain.
fn canonicalize_variables(&self) -> (Self, Bindings) {
fn go(v: &Value, bindings: &mut Bindings, counter: &mut usize) -> Value {
match v {
Value::Num(_) => v.clone(),
Value::Identifier(name) => {
match bindings.get(name) {
Some(id) => id.clone(),
None => {
let mut next_id = InlinableString::new();
write!(next_id, "{:X}", counter).unwrap();
let next_id = Value::Identifier(next_id);
*counter += 1;
bindings.insert(name.clone(), next_id.clone());
next_id
}
}
},
Value::Call(head, args) => {
Value::Call(head.clone(), args.iter().map(|x| go(x, bindings, counter)).collect())
}
}
}
let mut vars = HashMap::new();
let mut ctr = 0;
(go(self, &mut vars, &mut ctr), vars)
}
// Hash the canonical-variables form of a pattern. Allows patterns to be checked for equality safely.
fn pattern_hash(&self) -> u64 {
self.canonicalize_variables().0.get_hash()
}
// Generate all possible ways a pattern can be ordered, given commutative operators it may contain.
// This also recurses into child nodes.
pub fn pattern_reorderings(&self, env: &Env) -> Vec<Self> {
// Filter out redundant patterns from a result, and convert the argument lists back into Values
// Due to typing, this has to be factored out into a separate generic function
// rather than being part of the main logic below.
fn map_result<I: Iterator<Item=Vec<Value>>>(head: &InlinableString, result: I) -> Vec<Value> {
let mut existing_patterns = HashSet::new();
result.flat_map(|x| {
let resulting_value = Value::Call(head.clone(), x);
let hash = resulting_value.pattern_hash();
if existing_patterns.contains(&hash) {
None
} else {
existing_patterns.insert(hash);
Some(resulting_value)
}
}).collect()
}
match self {
// Call expressions can have their child nodes reordered and can be reordered themselves, if the head is a commutative operator
Value::Call(head, args) => {
let result = args.iter()
// Recursive step: generate all the valid reorderings of each child node.
.map(|x| x.pattern_reorderings(env))
// Generate all possible combinations of those reorderings.
.multi_cartesian_product();
// Generate all possible permutations of those combinations, if the operation allows for this.
if env.get_op(head).commutative {
map_result(head, result.flat_map(|comb| {
let k = comb.len();
comb.into_iter().permutations(k)
}))
} else {
map_result(head, result)
}
},
// Any other expression type is not reorderable.
_ => {
vec![self.clone()]
}
}
}
// Substitute bindings into an expression tree;
// the main match_and_bind function can also do this, but doing it here is more efficient
// when its full power is not needed
pub fn subst(&self, bindings: &Bindings) -> Value {
match self {
Value::Identifier(name) => {
match bindings.get(name) {
Some(value) => value.clone(),
None => Value::Identifier(name.clone())
}
},
Value::Call(fun, args) => Value::Call(fun.clone(), args.iter().map(|x| x.subst(bindings)).collect()),
x => x.clone()
}
}
// Ensure that a value is a number, returning an error otherwise.
pub fn assert_num(&self, ctx: &'static str) -> Result<i128> {
match self {
Value::Num(n) => Ok(*n),
_ => Err(anyhow!("expected number, got {:?}", self).context(ctx))
}
}
// The same but for identfiers.
pub fn assert_ident(&self, ctx: &'static str) -> Result<InlinableString> {
match self {
Value::Identifier(i) => Ok(i.clone()),
_ => Err(anyhow!("expected identifier, got {:?}", self).context(ctx))
}
}
pub fn render<W: fmt::Write>(&self, f: &mut W, env: &Env) -> fmt::Result {
fn go<W: fmt::Write>(v: &Value, parent_prec: Option<usize>, env: &Env, f: &mut W) -> fmt::Result {
match v {
// As unary - isn't parsed, negative numbers are written with Negate instead.
Value::Num(x) => if *x >= 0 {
write!(f, "{}", x)
} else { write!(f, "Negate[{}]", -x) },
Value::Identifier(i) => write!(f, "{}", i),
Value::Call(head, args) => {
match env.ops.get(head) {
Some(_) => {
// If the precedence of the enclosing operation is greater than or equal to this one's,
// add brackets around this one
let this_prec = precedence(head.chars().next().unwrap());
let render_brackets = match parent_prec {
Some(prec) => prec >= this_prec,
None => false
};
if render_brackets {
write!(f, "(")?;
}
for (i, arg) in args.iter().enumerate() {
go(arg, Some(this_prec), env, f)?;
if i + 1 != args.len() {
write!(f, "{}", head)?;
}
}
if render_brackets {
write!(f, ")")?;
}
},
// Just write a call expression with square brackets.
None => {
write!(f, "{}[", head)?;
for (i, arg) in args.iter().enumerate() {
go(arg, None, env, f)?;
if i + 1 != args.len() {
write!(f, ", ")?;
}
}
write!(f, "]")?;
}
}
Ok(())
}
}
}
go(self, None, env, f)
}
pub fn render_to_string(&self, env: &Env) -> InlinableString {
let mut out = InlinableString::new();
self.render(&mut out, env).unwrap();
out
}
}