diff --git a/benchmark/LD_disc_o.jl b/pr32_benchmarks/LD_disc_o.jl similarity index 98% rename from benchmark/LD_disc_o.jl rename to pr32_benchmarks/LD_disc_o.jl index 74a1668..0e5bbd6 100644 --- a/benchmark/LD_disc_o.jl +++ b/pr32_benchmarks/LD_disc_o.jl @@ -1,3 +1,5 @@ +#LightDark from POMDPModels.jl, modified for discrete binned observations + using Distributions # A one-dimensional light-dark problem, originally used to test MCVI # A very simple POMDP with continuous state and observation spaces. diff --git a/benchmark/Project.toml b/pr32_benchmarks/Project.toml similarity index 100% rename from benchmark/Project.toml rename to pr32_benchmarks/Project.toml diff --git a/benchmark/benchmarks.jl b/pr32_benchmarks/benchmarks.jl similarity index 90% rename from benchmark/benchmarks.jl rename to pr32_benchmarks/benchmarks.jl index 87c2ddc..4d4e603 100644 --- a/benchmark/benchmarks.jl +++ b/pr32_benchmarks/benchmarks.jl @@ -1,8 +1,12 @@ +#Benchmarks for ARDESPOT.jl PR No. 32 (Use OrderedDict) + +#Note: Uses HistoryRecorder instead of calls to action directly. This may add additional noise. + using Pkg Pkg.activate(".") using ARDESPOT -Pkg.activate("benchmark") +Pkg.activate("pr32_benchmarks") using BenchmarkTools using POMDPs diff --git a/benchmark/ld_policies.jl b/pr32_benchmarks/ld_policies.jl similarity index 97% rename from benchmark/ld_policies.jl rename to pr32_benchmarks/ld_policies.jl index a115b60..75e7a4d 100644 --- a/benchmark/ld_policies.jl +++ b/pr32_benchmarks/ld_policies.jl @@ -1,3 +1,4 @@ +#LD Policies for Bounds struct LDPolicy{M<:POMDP,S}<:Policy m::M particle_dict::Dict{S,Float64} @@ -11,7 +12,7 @@ function LDPolicy(m::LightDark1D{S}) where S return LDPolicy(m,Dict{S,Float64}(),-1,planner) end -##Tyler's map_bel +##Tyler's map_bel: @inline function get_incr!(h::Dict{K,V}, key::K, v) where {K,V} # modified from dict.jl index = Base.ht_keyindex2!(h, key)