11 Commits

Author SHA1 Message Date
387475d494 tokio-ify source-to-snf 2025-02-01 06:14:26 -08:00
e7aac2c796 tokio-ify roblox-to-snf 2025-01-31 13:11:49 -08:00
f13c7ddd48 from memory 2025-01-31 12:18:43 -08:00
97eebd3f8b update deps 2025-01-31 12:07:25 -08:00
2917fded43 roblox: convert textures during download 2025-01-31 12:07:25 -08:00
5c8a35fb20 change output folders 2025-01-30 13:33:35 -08:00
ee9b7fdc80 fix loader 2025-01-29 16:34:08 -08:00
c43bbd3410 fix loader 2025-01-29 14:53:35 -08:00
2ce8d4e2f8 decode in parallel, download one at a time
this will probably still hit the roblox rate limit
2025-01-27 15:18:30 -08:00
072adf1f87 download_assets with naive exponential backoff 2025-01-27 11:11:28 -08:00
db3ab1ec4b add rbx_asset + tokio + futures deps 2025-01-27 10:28:59 -08:00
7 changed files with 783 additions and 879 deletions

72
Cargo.lock generated
View File

@@ -698,9 +698,7 @@ checksum = "835a3dc7d1ec9e75e2b5fb4ba75396837112d2060b03f7d43bc1897c7f7211da"
[[package]] [[package]]
name = "fixed_wide" name = "fixed_wide"
version = "0.1.2" version = "0.1.1"
source = "sparse+https://git.itzana.me/api/packages/strafesnet/cargo/"
checksum = "a7e01a5b738e313c912fc41c425cf36e10c51647d3fd21d96db3d616344549fa"
dependencies = [ dependencies = [
"arrayvec", "arrayvec",
"bnum", "bnum",
@@ -1261,7 +1259,7 @@ dependencies = [
"image", "image",
"intel_tex_2", "intel_tex_2",
"strum", "strum",
"thiserror 1.0.69", "thiserror",
] ]
[[package]] [[package]]
@@ -1441,8 +1439,6 @@ checksum = "8355be11b20d696c8f18f6cc018c4e372165b1fa8126cef092399c9951984ffa"
[[package]] [[package]]
name = "linear_ops" name = "linear_ops"
version = "0.1.0" version = "0.1.0"
source = "sparse+https://git.itzana.me/api/packages/strafesnet/cargo/"
checksum = "b2e6977ac24f47086d8a7a2d4ae1c720e86dfdc8407cf5e34c18bfa01053c456"
dependencies = [ dependencies = [
"fixed_wide", "fixed_wide",
"paste", "paste",
@@ -1568,7 +1564,7 @@ dependencies = [
[[package]] [[package]]
name = "map-tool" name = "map-tool"
version = "2.0.0" version = "1.6.0"
dependencies = [ dependencies = [
"anyhow", "anyhow",
"clap", "clap",
@@ -1587,7 +1583,6 @@ dependencies = [
"strafesnet_deferred_loader", "strafesnet_deferred_loader",
"strafesnet_rbx_loader", "strafesnet_rbx_loader",
"strafesnet_snf", "strafesnet_snf",
"thiserror 2.0.11",
"tokio", "tokio",
"vbsp", "vbsp",
"vmdl", "vmdl",
@@ -1620,7 +1615,7 @@ checksum = "317f146e2eb7021892722af37cf1b971f0a70c8406f487e24952667616192c64"
dependencies = [ dependencies = [
"cfg-if", "cfg-if",
"miette-derive", "miette-derive",
"thiserror 1.0.69", "thiserror",
"unicode-width", "unicode-width",
] ]
@@ -2175,8 +2170,6 @@ dependencies = [
[[package]] [[package]]
name = "ratio_ops" name = "ratio_ops"
version = "0.1.0" version = "0.1.0"
source = "sparse+https://git.itzana.me/api/packages/strafesnet/cargo/"
checksum = "01239195d6afe0509e7e3511b716c0540251dfe7ece0a9a5a27116afb766c42c"
[[package]] [[package]]
name = "rav1e" name = "rav1e"
@@ -2208,7 +2201,7 @@ dependencies = [
"rand_chacha", "rand_chacha",
"simd_helpers", "simd_helpers",
"system-deps", "system-deps",
"thiserror 1.0.69", "thiserror",
"v_frame", "v_frame",
"wasm-bindgen", "wasm-bindgen",
] ]
@@ -2274,7 +2267,7 @@ dependencies = [
"rbx_dom_weak", "rbx_dom_weak",
"rbx_reflection", "rbx_reflection",
"rbx_reflection_database", "rbx_reflection_database",
"thiserror 1.0.69", "thiserror",
] ]
[[package]] [[package]]
@@ -2305,7 +2298,7 @@ checksum = "c1b43fe592a4ce6fe54eb215fb82735efbb516d2cc045a94e3dc0234ff293620"
dependencies = [ dependencies = [
"rbx_types", "rbx_types",
"serde", "serde",
"thiserror 1.0.69", "thiserror",
] ]
[[package]] [[package]]
@@ -2332,7 +2325,7 @@ dependencies = [
"lazy_static", "lazy_static",
"rand", "rand",
"serde", "serde",
"thiserror 1.0.69", "thiserror",
] ]
[[package]] [[package]]
@@ -2352,8 +2345,6 @@ dependencies = [
[[package]] [[package]]
name = "rbxassetid" name = "rbxassetid"
version = "0.1.0" version = "0.1.0"
source = "sparse+https://git.itzana.me/api/packages/strafesnet/cargo/"
checksum = "e6821fe9eaff54cd142932cb04c612b7599d9b8586973145b7ec1230ae84d184"
dependencies = [ dependencies = [
"url", "url",
] ]
@@ -2487,8 +2478,6 @@ dependencies = [
[[package]] [[package]]
name = "roblox_emulator" name = "roblox_emulator"
version = "0.4.7" version = "0.4.7"
source = "sparse+https://git.itzana.me/api/packages/strafesnet/cargo/"
checksum = "7fc98335ce4b8548b725d727c5b32bd0b38274606c48fce6b6e7e5807d94db6b"
dependencies = [ dependencies = [
"glam", "glam",
"mlua", "mlua",
@@ -2750,22 +2739,17 @@ checksum = "a2eb9349b6444b326872e140eb1cf5e7c522154d69e7a0ffb0fb81c06b37543f"
[[package]] [[package]]
name = "strafesnet_bsp_loader" name = "strafesnet_bsp_loader"
version = "0.3.0" version = "0.3.0"
source = "sparse+https://git.itzana.me/api/packages/strafesnet/cargo/"
checksum = "3cc98773f2b98eb708b098946870b769975b63a396b84698b67e3d968029005d"
dependencies = [ dependencies = [
"glam", "glam",
"strafesnet_common", "strafesnet_common",
"strafesnet_deferred_loader", "strafesnet_deferred_loader",
"vbsp", "vbsp",
"vmdl", "vmdl",
"vpk",
] ]
[[package]] [[package]]
name = "strafesnet_common" name = "strafesnet_common"
version = "0.6.0" version = "0.6.0"
source = "sparse+https://git.itzana.me/api/packages/strafesnet/cargo/"
checksum = "0c1d7a83e1f6b579c6a9b4dc70c92373ab53b938601cd75928dd6795b5ffef21"
dependencies = [ dependencies = [
"arrayvec", "arrayvec",
"bitflags 2.8.0", "bitflags 2.8.0",
@@ -2779,8 +2763,6 @@ dependencies = [
[[package]] [[package]]
name = "strafesnet_deferred_loader" name = "strafesnet_deferred_loader"
version = "0.5.0" version = "0.5.0"
source = "sparse+https://git.itzana.me/api/packages/strafesnet/cargo/"
checksum = "63d5d48e587d5f8bf5385bee3505ed790727fef68de855cf58247a08c5952bef"
dependencies = [ dependencies = [
"strafesnet_common", "strafesnet_common",
] ]
@@ -2788,8 +2770,6 @@ dependencies = [
[[package]] [[package]]
name = "strafesnet_rbx_loader" name = "strafesnet_rbx_loader"
version = "0.6.0" version = "0.6.0"
source = "sparse+https://git.itzana.me/api/packages/strafesnet/cargo/"
checksum = "e4659a49128c8d12b9fbdb289969cae04bfc5c1750d4273897700c5c17730d8a"
dependencies = [ dependencies = [
"bytemuck", "bytemuck",
"glam", "glam",
@@ -2807,9 +2787,7 @@ dependencies = [
[[package]] [[package]]
name = "strafesnet_snf" name = "strafesnet_snf"
version = "0.3.0" version = "0.2.0"
source = "sparse+https://git.itzana.me/api/packages/strafesnet/cargo/"
checksum = "fd24a22c484ca04213fa44b1d34bfbec385f0d176a2b5829cfa59ba7987b80d5"
dependencies = [ dependencies = [
"binrw 0.14.1", "binrw 0.14.1",
"id", "id",
@@ -2996,16 +2974,7 @@ version = "1.0.69"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "b6aaf5339b578ea85b50e080feb250a3e8ae8cfcdff9a461c9ec2904bc923f52" checksum = "b6aaf5339b578ea85b50e080feb250a3e8ae8cfcdff9a461c9ec2904bc923f52"
dependencies = [ dependencies = [
"thiserror-impl 1.0.69", "thiserror-impl",
]
[[package]]
name = "thiserror"
version = "2.0.11"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "d452f284b73e6d76dd36758a0c8684b1d5be31f92b89d07fd5822175732206fc"
dependencies = [
"thiserror-impl 2.0.11",
] ]
[[package]] [[package]]
@@ -3019,17 +2988,6 @@ dependencies = [
"syn 2.0.96", "syn 2.0.96",
] ]
[[package]]
name = "thiserror-impl"
version = "2.0.11"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "26afc1baea8a989337eeb52b6e72a039780ce45c3edfcc9c5b9d112feeb173c2"
dependencies = [
"proc-macro2",
"quote",
"syn 2.0.96",
]
[[package]] [[package]]
name = "tiff" name = "tiff"
version = "0.9.1" version = "0.9.1"
@@ -3296,7 +3254,7 @@ dependencies = [
"num_enum", "num_enum",
"serde", "serde",
"static_assertions", "static_assertions",
"thiserror 1.0.69", "thiserror",
"vdf-reader", "vdf-reader",
"zip-lzma", "zip-lzma",
] ]
@@ -3317,7 +3275,7 @@ dependencies = [
"miette", "miette",
"parse-display 0.9.1", "parse-display 0.9.1",
"serde", "serde",
"thiserror 1.0.69", "thiserror",
] ]
[[package]] [[package]]
@@ -3346,7 +3304,7 @@ dependencies = [
"itertools 0.13.0", "itertools 0.13.0",
"num_enum", "num_enum",
"static_assertions", "static_assertions",
"thiserror 1.0.69", "thiserror",
"tracing", "tracing",
] ]
@@ -3359,7 +3317,7 @@ dependencies = [
"miette", "miette",
"serde", "serde",
"serde_repr", "serde_repr",
"thiserror 1.0.69", "thiserror",
"vdf-reader", "vdf-reader",
] ]
@@ -3371,7 +3329,7 @@ checksum = "60ec10e731515f58d5494d472f027d9c6fc8500fcb790ff55751031bcad87b6b"
dependencies = [ dependencies = [
"ahash", "ahash",
"binrw 0.13.3", "binrw 0.13.3",
"thiserror 1.0.69", "thiserror",
] ]
[[package]] [[package]]

View File

@@ -1,25 +1,13 @@
[package] [package]
name = "map-tool" name = "map-tool"
version = "2.0.0" version = "1.6.0"
edition = "2021" edition = "2021"
# See more keys and their definitions at https://doc.rust-lang.org/cargo/reference/manifest.html # See more keys and their definitions at https://doc.rust-lang.org/cargo/reference/manifest.html
[lib]
name = "map_tool"
path = "src/lib.rs"
[[bin]]
name = "map-tool"
path = "src/main.rs"
required-features = ["cli"]
[features]
cli = ["dep:clap"]
[dependencies] [dependencies]
anyhow = "1.0.75" anyhow = "1.0.75"
clap = { version = "4.4.2", features = ["derive"], optional = true } clap = { version = "4.4.2", features = ["derive"] }
flate2 = "1.0.27" flate2 = "1.0.27"
futures = "0.3.31" futures = "0.3.31"
image = "0.25.2" image = "0.25.2"
@@ -30,12 +18,10 @@ rbx_binary = { version = "0.7.4", registry = "strafesnet" }
rbx_dom_weak = { version = "2.7.0", registry = "strafesnet" } rbx_dom_weak = { version = "2.7.0", registry = "strafesnet" }
rbx_reflection_database = { version = "0.2.10", registry = "strafesnet" } rbx_reflection_database = { version = "0.2.10", registry = "strafesnet" }
rbx_xml = { version = "0.13.3", registry = "strafesnet" } rbx_xml = { version = "0.13.3", registry = "strafesnet" }
rbxassetid = { version = "0.1.0", registry = "strafesnet" } strafesnet_bsp_loader = { version = "0.2.1", registry = "strafesnet" }
strafesnet_bsp_loader = { version = "0.3.0", registry = "strafesnet" } strafesnet_deferred_loader = { version = "0.4.0", features = ["legacy"], registry = "strafesnet" }
strafesnet_deferred_loader = { version = "0.5.0", registry = "strafesnet" } strafesnet_rbx_loader = { version = "0.5.1", registry = "strafesnet" }
strafesnet_rbx_loader = { version = "0.6.0", registry = "strafesnet" } strafesnet_snf = { version = "0.2.0", registry = "strafesnet" }
strafesnet_snf = { version = "0.3.0", registry = "strafesnet" }
thiserror = "2.0.11"
tokio = { version = "1.43.0", features = ["macros", "rt-multi-thread", "fs"] } tokio = { version = "1.43.0", features = ["macros", "rt-multi-thread", "fs"] }
vbsp = "0.6.0" vbsp = "0.6.0"
vmdl = "0.2.0" vmdl = "0.2.0"

75
src/common.rs Normal file
View File

@@ -0,0 +1,75 @@
use std::path::PathBuf;
use std::io::{Read,Seek};
use anyhow::Result as AResult;
fn load_image<R:Read+Seek+std::io::BufRead>(input:&mut R)->AResult<image::DynamicImage>{
let mut fourcc=[0u8;4];
input.read_exact(&mut fourcc)?;
input.rewind()?;
match &fourcc{
b"\x89PNG"=>Ok(image::load(input,image::ImageFormat::Png)?),
b"\xFF\xD8\xFF\xE0"=>Ok(image::load(input,image::ImageFormat::Jpeg)?),//JFIF
b"<rob"=>Err(anyhow::Error::msg("Roblox xml garbage is not supported yet")),
other=>Err(anyhow::Error::msg(format!("Unknown texture format {:?}",other))),
}
}
fn convert(file_thing:std::fs::DirEntry) -> AResult<()>{
let mut input = std::io::BufReader::new(std::fs::File::open(file_thing.path())?);
let image=load_image(&mut input)?.to_rgba8();//this sets a=255, arcane is actually supposed to look like that
let format=if image.width()%4!=0||image.height()%4!=0{
image_dds::ImageFormat::Rgba8UnormSrgb
}else{
image_dds::ImageFormat::BC7RgbaUnormSrgb
};
//this fails if the image dimensions are not a multiple of 4
let dds = image_dds::dds_from_image(
&image,
format,
image_dds::Quality::Slow,
image_dds::Mipmaps::GeneratedAutomatic,
)?;
//write dds
let mut dest=PathBuf::from("textures");
dest.push(file_thing.file_name());
dest.set_extension("dds");
let mut writer = std::io::BufWriter::new(std::fs::File::create(dest)?);
dds.write(&mut writer)?;
//move file to processed
let mut dest=PathBuf::from("textures/processed");
dest.push(file_thing.file_name());
std::fs::rename(file_thing.path(), dest)?;
Ok(())
}
pub fn convert_textures() -> AResult<()>{
std::fs::create_dir_all("textures/unprocessed")?;
std::fs::create_dir_all("textures/processed")?;
let start = std::time::Instant::now();
let mut threads=Vec::new();
for entry in std::fs::read_dir("textures/unprocessed")? {
let file_thing=entry?;
threads.push(std::thread::spawn(move ||{
let file_name=format!("{:?}",file_thing);
let result=convert(file_thing);
if let Err(e)=result{
println!("error processing file:{:?} error message:{:?}",file_name,e);
}
}));
}
let mut i=0;
let n_threads=threads.len();
for thread in threads{
i+=1;
if let Err(e)=thread.join(){
println!("thread error: {:?}",e);
}else{
println!("{}/{}",i,n_threads);
}
}
println!("{:?}", start.elapsed());
Ok(())
}

View File

@@ -1,2 +0,0 @@
pub mod roblox;
pub mod source;

View File

@@ -1,3 +1,7 @@
mod common;
mod roblox;
mod source;
use clap::{Parser,Subcommand}; use clap::{Parser,Subcommand};
use anyhow::Result as AResult; use anyhow::Result as AResult;
@@ -12,9 +16,10 @@ struct Cli {
#[derive(Subcommand)] #[derive(Subcommand)]
enum Commands{ enum Commands{
#[command(flatten)] #[command(flatten)]
Roblox(map_tool::roblox::Commands), Roblox(roblox::Commands),
#[command(flatten)] #[command(flatten)]
Source(map_tool::source::Commands), Source(source::Commands),
ConvertTextures,
} }
#[tokio::main] #[tokio::main]
@@ -23,5 +28,6 @@ async fn main()->AResult<()>{
match cli.command{ match cli.command{
Commands::Roblox(commands)=>commands.run().await, Commands::Roblox(commands)=>commands.run().await,
Commands::Source(commands)=>commands.run().await, Commands::Source(commands)=>commands.run().await,
Commands::ConvertTextures=>common::convert_textures(),
} }
} }

View File

@@ -1,40 +1,60 @@
use std::path::{Path,PathBuf};
use std::io::{Cursor,Read,Seek}; use std::io::{Cursor,Read,Seek};
use std::collections::HashSet; use std::collections::HashSet;
use clap::{Args,Subcommand};
use anyhow::Result as AResult;
use rbx_dom_weak::Instance; use rbx_dom_weak::Instance;
use strafesnet_deferred_loader::deferred_loader::LoadFailureMode; use strafesnet_deferred_loader::deferred_loader::LoadFailureMode;
use rbxassetid::RobloxAssetId; use rbxassetid::RobloxAssetId;
use tokio::io::AsyncReadExt;
// === Public library API === const DOWNLOAD_LIMIT:usize=16;
/// Unique asset IDs referenced by a Roblox place/model file. #[derive(Subcommand)]
#[derive(Default)] pub enum Commands{
pub struct UniqueAssets{ RobloxToSNF(RobloxToSNFSubcommand),
pub meshes:HashSet<RobloxAssetId>, DownloadAssets(DownloadAssetsSubcommand),
pub unions:HashSet<RobloxAssetId>,
pub textures:HashSet<RobloxAssetId>,
} }
#[derive(Debug,thiserror::Error)] #[derive(Args)]
pub enum LoadDomError{ pub struct RobloxToSNFSubcommand {
#[error("IO error {0:?}")] #[arg(long)]
IO(#[from]std::io::Error), output_folder:PathBuf,
#[error("Binary decode error {0:?}")] #[arg(required=true)]
input_files:Vec<PathBuf>,
}
#[derive(Args)]
pub struct DownloadAssetsSubcommand{
#[arg(required=true)]
roblox_files:Vec<PathBuf>,
// #[arg(long)]
// cookie_file:Option<String>,
}
impl Commands{
pub async fn run(self)->AResult<()>{
match self{
Commands::RobloxToSNF(subcommand)=>roblox_to_snf(subcommand.input_files,subcommand.output_folder).await,
Commands::DownloadAssets(subcommand)=>download_assets(
subcommand.roblox_files,
rbx_asset::cookie::Cookie::new("".to_string()),
).await,
}
}
}
#[allow(unused)]
#[derive(Debug)]
enum LoadDomError{
IO(std::io::Error),
Binary(rbx_binary::DecodeError), Binary(rbx_binary::DecodeError),
#[error("XML decode error {0:?}")]
Xml(rbx_xml::DecodeError), Xml(rbx_xml::DecodeError),
#[error("Unknown file format")]
UnknownFormat, UnknownFormat,
} }
fn load_dom<R:Read+Seek>(mut input:R)->Result<rbx_dom_weak::WeakDom,LoadDomError>{
/// Parse a Roblox file (binary or XML) from bytes into a WeakDom.
pub fn load_dom(data:&[u8])->Result<rbx_dom_weak::WeakDom,LoadDomError>{
load_dom_reader(Cursor::new(data))
}
fn load_dom_reader<R:Read+Seek>(mut input:R)->Result<rbx_dom_weak::WeakDom,LoadDomError>{
let mut first_8=[0u8;8]; let mut first_8=[0u8;8];
input.read_exact(&mut first_8)?; input.read_exact(&mut first_8).map_err(LoadDomError::IO)?;
input.rewind()?; input.rewind().map_err(LoadDomError::IO)?;
match &first_8{ match &first_8{
b"<roblox!"=>rbx_binary::from_reader(input).map_err(LoadDomError::Binary), b"<roblox!"=>rbx_binary::from_reader(input).map_err(LoadDomError::Binary),
b"<roblox "=>rbx_xml::from_reader(input,rbx_xml::DecodeOptions::default()).map_err(LoadDomError::Xml), b"<roblox "=>rbx_xml::from_reader(input,rbx_xml::DecodeOptions::default()).map_err(LoadDomError::Xml),
@@ -42,113 +62,56 @@ fn load_dom_reader<R:Read+Seek>(mut input:R)->Result<rbx_dom_weak::WeakDom,LoadD
} }
} }
/// Scan a parsed DOM and return all unique asset IDs (meshes, textures, unions). /* The ones I'm interested in:
pub fn get_unique_assets(dom:rbx_dom_weak::WeakDom)->UniqueAssets{ Beam.Texture
let mut assets=UniqueAssets::default(); Decal.Texture
for object in dom.into_raw().1.into_values(){ FileMesh.MeshId
assets.collect(&object); FileMesh.TextureId
} MaterialVariant.ColorMap
assets MaterialVariant.MetalnessMap
} MaterialVariant.NormalMap
MaterialVariant.RoughnessMap
/// Scan a Roblox file (bytes) and return all unique asset IDs. MeshPart.MeshId
pub fn get_unique_assets_from_file(data:&[u8])->Result<UniqueAssets,LoadDomError>{ MeshPart.TextureID
let dom=load_dom(data)?; ParticleEmitter.Texture
Ok(get_unique_assets(dom)) Sky.MoonTextureId
} Sky.SkyboxBk
Sky.SkyboxDn
#[derive(Debug,thiserror::Error)] Sky.SkyboxFt
pub enum ConvertTextureError{ Sky.SkyboxLf
#[error("Image error {0:?}")] Sky.SkyboxRt
Image(#[from]image::ImageError), Sky.SkyboxUp
#[error("DDS create error {0:?}")] Sky.SunTextureId
DDS(#[from]image_dds::CreateDdsError), SurfaceAppearance.ColorMap
#[error("DDS write error {0:?}")] SurfaceAppearance.MetalnessMap
DDSWrite(#[from]image_dds::ddsfile::Error), SurfaceAppearance.NormalMap
} SurfaceAppearance.RoughnessMap
SurfaceAppearance.TexturePack
/// Convert image bytes (PNG, JPEG, etc.) into DDS texture bytes. */
pub fn convert_texture_to_dds(image_data:&[u8])->Result<Vec<u8>,ConvertTextureError>{ fn accumulate_content_id(content_list:&mut HashSet<RobloxAssetId>,object:&Instance,property:&str){
let image=image::load_from_memory(image_data)?.to_rgba8(); if let Some(rbx_dom_weak::types::Variant::Content(content))=object.properties.get(property){
let url:&str=content.as_ref();
let format=if image.width()%4!=0||image.height()%4!=0{ if let Ok(asset_id)=url.parse(){
image_dds::ImageFormat::Rgba8UnormSrgb content_list.insert(asset_id);
}else{ }else{
image_dds::ImageFormat::BC7RgbaUnormSrgb println!("Content failed to parse into AssetID: {:?}",content);
}; }
let dds=image_dds::dds_from_image(
&image,
format,
image_dds::Quality::Slow,
image_dds::Mipmaps::GeneratedAutomatic,
)?;
let mut buf=Vec::new();
dds.write(&mut Cursor::new(&mut buf))?;
Ok(buf)
}
#[derive(Debug,thiserror::Error)]
pub enum ConvertError{
#[error("IO error {0:?}")]
IO(#[from]std::io::Error),
#[error("SNF map error {0:?}")]
SNFMap(strafesnet_snf::map::Error),
#[error("Roblox read error {0:?}")]
RobloxRead(strafesnet_rbx_loader::ReadError),
#[error("Roblox load error {0:?}")]
RobloxLoad(strafesnet_rbx_loader::LoadError),
}
/// Convert a Roblox place/model file (bytes) to SNF map format (bytes).
pub fn convert_to_snf(data:&[u8])->Result<Vec<u8>,ConvertError>{
let model=strafesnet_rbx_loader::read(
Cursor::new(data)
).map_err(ConvertError::RobloxRead)?;
let mut place=model.into_place();
place.run_scripts();
let map=place.to_snf(LoadFailureMode::DefaultToNone).map_err(ConvertError::RobloxLoad)?;
let mut buf=Vec::new();
strafesnet_snf::map::write_map(Cursor::new(&mut buf),map).map_err(ConvertError::SNFMap)?;
Ok(buf)
}
/// Download a single asset from Roblox by ID. Returns raw asset bytes.
pub async fn download_asset(context:&rbx_asset::cookie::CookieContext,asset_id:u64)->Result<Vec<u8>,rbx_asset::cookie::GetError>{
context.get_asset(rbx_asset::cookie::GetAssetRequest{
asset_id,
version:None,
}).await
}
/// Download a single asset with retry and exponential backoff for rate limiting.
/// Returns None if all retries are exhausted or a non-rate-limit error occurs.
pub async fn download_asset_retry(context:&rbx_asset::cookie::CookieContext,asset_id:u64)->Option<Vec<u8>>{
const BACKOFF_MUL:f32=1.3956124250860895286;
let mut backoff=1000f32;
for _ in 0..12{
match download_asset(context,asset_id).await{
Ok(data)=>return Some(data),
Err(rbx_asset::cookie::GetError::Response(rbx_asset::ResponseError::StatusCodeWithUrlAndBody(scwuab)))=>{
if scwuab.status_code.as_u16()==429{
tokio::time::sleep(std::time::Duration::from_millis(backoff as u64)).await;
backoff*=BACKOFF_MUL;
}else{ }else{
return None; println!("property={} does not exist for class={}",property,object.class.as_str());
} }
},
Err(_)=>return None,
}
}
None
} }
async fn read_entire_file(path:impl AsRef<Path>)->Result<Cursor<Vec<u8>>,std::io::Error>{
// === Private helpers === let mut file=tokio::fs::File::open(path).await?;
let mut data=Vec::new();
file.read_to_end(&mut data).await?;
Ok(Cursor::new(data))
}
#[derive(Default)]
struct UniqueAssets{
meshes:HashSet<RobloxAssetId>,
unions:HashSet<RobloxAssetId>,
textures:HashSet<RobloxAssetId>,
}
impl UniqueAssets{ impl UniqueAssets{
fn collect(&mut self,object:&Instance){ fn collect(&mut self,object:&Instance){
match object.class.as_str(){ match object.class.as_str(){
@@ -178,75 +141,28 @@ impl UniqueAssets{
} }
} }
fn accumulate_content_id(content_list:&mut HashSet<RobloxAssetId>,object:&Instance,property:&str){ #[allow(unused)]
if let Some(rbx_dom_weak::types::Variant::Content(content))=object.properties.get(property){ #[derive(Debug)]
let url:&str=content.as_ref(); enum UniqueAssetError{
if let Ok(asset_id)=url.parse(){ IO(std::io::Error),
content_list.insert(asset_id); LoadDom(LoadDomError),
}else{
println!("Content failed to parse into AssetID: {:?}",content);
}
}else{
println!("property={} does not exist for class={}",property,object.class.as_str());
}
} }
async fn unique_assets(path:&Path)->Result<UniqueAssets,UniqueAssetError>{
// === CLI === // read entire file
let mut assets=UniqueAssets::default();
#[cfg(feature="cli")] let data=read_entire_file(path).await.map_err(UniqueAssetError::IO)?;
mod cli{ let dom=load_dom(data).map_err(UniqueAssetError::LoadDom)?;
use super::*; for object in dom.into_raw().1.into_values(){
use std::path::{Path,PathBuf}; assets.collect(&object);
use clap::{Args,Subcommand};
use anyhow::Result as AResult;
use tokio::io::AsyncReadExt;
const DOWNLOAD_LIMIT:usize=16;
#[derive(Subcommand)]
pub enum Commands{
RobloxToSNF(RobloxToSNFSubcommand),
DownloadAssets(DownloadAssetsSubcommand),
} }
Ok(assets)
#[derive(Args)] }
pub struct RobloxToSNFSubcommand { enum DownloadType{
#[arg(long)]
output_folder:PathBuf,
#[arg(required=true)]
input_files:Vec<PathBuf>,
}
#[derive(Args)]
pub struct DownloadAssetsSubcommand{
#[arg(required=true)]
roblox_files:Vec<PathBuf>,
}
impl Commands{
pub async fn run(self)->AResult<()>{
match self{
Commands::RobloxToSNF(subcommand)=>cli_roblox_to_snf(subcommand.input_files,subcommand.output_folder).await,
Commands::DownloadAssets(subcommand)=>cli_download_assets(
subcommand.roblox_files,
rbx_asset::cookie::Cookie::new("".to_string()),
).await,
}
}
}
async fn read_entire_file(path:impl AsRef<Path>)->Result<Vec<u8>,std::io::Error>{
let mut file=tokio::fs::File::open(path).await?;
let mut data=Vec::new();
file.read_to_end(&mut data).await?;
Ok(data)
}
enum DownloadType{
Texture(RobloxAssetId), Texture(RobloxAssetId),
Mesh(RobloxAssetId), Mesh(RobloxAssetId),
Union(RobloxAssetId), Union(RobloxAssetId),
} }
impl DownloadType{ impl DownloadType{
fn path(&self)->PathBuf{ fn path(&self)->PathBuf{
match self{ match self{
DownloadType::Texture(asset_id)=>format!("downloaded_textures/{}",asset_id.0.to_string()).into(), DownloadType::Texture(asset_id)=>format!("downloaded_textures/{}",asset_id.0.to_string()).into(),
@@ -261,30 +177,33 @@ mod cli{
DownloadType::Union(asset_id)=>asset_id.0, DownloadType::Union(asset_id)=>asset_id.0,
} }
} }
} }
enum DownloadResult{ enum DownloadResult{
Cached(PathBuf), Cached(PathBuf),
Data(Vec<u8>), Data(Vec<u8>),
Failed, Failed,
} }
#[derive(Default,Debug)] #[derive(Default,Debug)]
struct Stats{ struct Stats{
total_assets:u32, total_assets:u32,
cached_assets:u32, cached_assets:u32,
downloaded_assets:u32, downloaded_assets:u32,
failed_downloads:u32, failed_downloads:u32,
timed_out_downloads:u32, timed_out_downloads:u32,
} }
async fn download_retry(stats:&mut Stats,context:&rbx_asset::cookie::CookieContext,download_instruction:DownloadType)->Result<DownloadResult,std::io::Error>{ async fn download_retry(stats:&mut Stats,context:&rbx_asset::cookie::CookieContext,download_instruction:DownloadType)->Result<DownloadResult,std::io::Error>{
stats.total_assets+=1; stats.total_assets+=1;
let download_instruction=download_instruction;
// check if file exists on disk
let path=download_instruction.path(); let path=download_instruction.path();
if tokio::fs::try_exists(path.as_path()).await?{ if tokio::fs::try_exists(path.as_path()).await?{
stats.cached_assets+=1; stats.cached_assets+=1;
return Ok(DownloadResult::Cached(path)); return Ok(DownloadResult::Cached(path));
} }
let asset_id=download_instruction.asset_id(); let asset_id=download_instruction.asset_id();
// if not, download file
let mut retry=0; let mut retry=0;
const BACKOFF_MUL:f32=1.3956124250860895286; const BACKOFF_MUL:f32=1.3956124250860895286;//exp(1/3)
let mut backoff=1000f32; let mut backoff=1000f32;
loop{ loop{
let asset_result=context.get_asset(rbx_asset::cookie::GetAssetRequest{ let asset_result=context.get_asset(rbx_asset::cookie::GetAssetRequest{
@@ -321,58 +240,109 @@ mod cli{
}, },
} }
} }
}
#[allow(unused)]
#[derive(Debug)]
enum ConvertTextureError{
Io(std::io::Error),
Image(image::ImageError),
DDS(image_dds::CreateDdsError),
DDSWrite(image_dds::ddsfile::Error),
}
impl From<std::io::Error> for ConvertTextureError{
fn from(value:std::io::Error)->Self{
Self::Io(value)
} }
}
async fn cli_convert_texture(asset_id:RobloxAssetId,download_result:DownloadResult)->Result<(),CliConvertTextureError>{ impl From<image::ImageError> for ConvertTextureError{
fn from(value:image::ImageError)->Self{
Self::Image(value)
}
}
impl From<image_dds::CreateDdsError> for ConvertTextureError{
fn from(value:image_dds::CreateDdsError)->Self{
Self::DDS(value)
}
}
impl From<image_dds::ddsfile::Error> for ConvertTextureError{
fn from(value:image_dds::ddsfile::Error)->Self{
Self::DDSWrite(value)
}
}
async fn convert_texture(asset_id:RobloxAssetId,download_result:DownloadResult)->Result<(),ConvertTextureError>{
let data=match download_result{ let data=match download_result{
DownloadResult::Cached(path)=>tokio::fs::read(path).await?, DownloadResult::Cached(path)=>tokio::fs::read(path).await?,
DownloadResult::Data(data)=>data, DownloadResult::Data(data)=>data,
DownloadResult::Failed=>return Ok(()), DownloadResult::Failed=>return Ok(()),
}; };
let dds_data=convert_texture_to_dds(&data)?; // image::ImageFormat::Png
// image::ImageFormat::Jpeg
let image=image::load_from_memory(&data)?.to_rgba8();
// pick format
let format=if image.width()%4!=0||image.height()%4!=0{
image_dds::ImageFormat::Rgba8UnormSrgb
}else{
image_dds::ImageFormat::BC7RgbaUnormSrgb
};
//this fails if the image dimensions are not a multiple of 4
let dds=image_dds::dds_from_image(
&image,
format,
image_dds::Quality::Slow,
image_dds::Mipmaps::GeneratedAutomatic,
)?;
let file_name=format!("textures/{}.dds",asset_id.0); let file_name=format!("textures/{}.dds",asset_id.0);
tokio::fs::write(file_name,dds_data).await?; let mut file=std::fs::File::create(file_name)?;
dds.write(&mut file)?;
Ok(()) Ok(())
} }
async fn download_assets(paths:Vec<PathBuf>,cookie:rbx_asset::cookie::Cookie)->AResult<()>{
#[derive(Debug,thiserror::Error)]
enum CliConvertTextureError{
#[error("IO error {0:?}")]
Io(#[from]std::io::Error),
#[error("Convert texture error {0:?}")]
Convert(#[from]ConvertTextureError),
}
async fn cli_download_assets(paths:Vec<PathBuf>,cookie:rbx_asset::cookie::Cookie)->AResult<()>{
tokio::try_join!( tokio::try_join!(
tokio::fs::create_dir_all("downloaded_textures"), tokio::fs::create_dir_all("downloaded_textures"),
tokio::fs::create_dir_all("textures"), tokio::fs::create_dir_all("textures"),
tokio::fs::create_dir_all("meshes"), tokio::fs::create_dir_all("meshes"),
tokio::fs::create_dir_all("unions"), tokio::fs::create_dir_all("unions"),
)?; )?;
// use mpsc
let thread_limit=std::thread::available_parallelism()?.get(); let thread_limit=std::thread::available_parallelism()?.get();
let (send_assets,mut recv_assets)=tokio::sync::mpsc::channel(DOWNLOAD_LIMIT); let (send_assets,mut recv_assets)=tokio::sync::mpsc::channel(DOWNLOAD_LIMIT);
let (send_texture,mut recv_texture)=tokio::sync::mpsc::channel(thread_limit); let (send_texture,mut recv_texture)=tokio::sync::mpsc::channel(thread_limit);
// map decode dispatcher
// read files multithreaded
// produce UniqueAssetsResult per file
tokio::spawn(async move{ tokio::spawn(async move{
// move send so it gets dropped when all maps have been decoded
// closing the channel
let mut it=paths.into_iter(); let mut it=paths.into_iter();
static SEM:tokio::sync::Semaphore=tokio::sync::Semaphore::const_new(0); static SEM:tokio::sync::Semaphore=tokio::sync::Semaphore::const_new(0);
SEM.add_permits(thread_limit); SEM.add_permits(thread_limit);
while let (Ok(permit),Some(path))=(SEM.acquire().await,it.next()){ while let (Ok(permit),Some(path))=(SEM.acquire().await,it.next()){
let send=send_assets.clone(); let send=send_assets.clone();
tokio::spawn(async move{ tokio::spawn(async move{
let data=read_entire_file(path.as_path()).await; let result=unique_assets(path.as_path()).await;
let result=data.map_err(LoadDomError::from).and_then(|d|{
let dom=load_dom(&d)?;
Ok(get_unique_assets(dom))
});
_=send.send(result).await; _=send.send(result).await;
drop(permit); drop(permit);
}); });
} }
}); });
// download manager
// insert into global unique assets guy
// add to download queue if the asset is globally unique and does not already exist on disk
let mut stats=Stats::default(); let mut stats=Stats::default();
let context=rbx_asset::cookie::CookieContext::new(cookie); let context=rbx_asset::cookie::CookieContext::new(cookie);
let mut globally_unique_assets=UniqueAssets::default(); let mut globally_unique_assets=UniqueAssets::default();
// pop a job = retry_queue.pop_front() or ingest(recv.recv().await)
// SLOW MODE:
// acquire all permits
// drop all permits
// pop one job
// if it succeeds go into fast mode
// FAST MODE:
// acquire one permit
// pop a job
let download_thread=tokio::spawn(async move{ let download_thread=tokio::spawn(async move{
while let Some(result)=recv_assets.recv().await{ while let Some(result)=recv_assets.recv().await{
let unique_assets=match result{ let unique_assets=match result{
@@ -406,7 +376,7 @@ mod cli{
SEM.add_permits(thread_limit); SEM.add_permits(thread_limit);
while let (Ok(permit),Some((asset_id,download_result)))=(SEM.acquire().await,recv_texture.recv().await){ while let (Ok(permit),Some((asset_id,download_result)))=(SEM.acquire().await,recv_texture.recv().await){
tokio::spawn(async move{ tokio::spawn(async move{
let result=cli_convert_texture(asset_id,download_result).await; let result=convert_texture(asset_id,download_result).await;
drop(permit); drop(permit);
result.unwrap(); result.unwrap();
}); });
@@ -414,9 +384,45 @@ mod cli{
download_thread.await??; download_thread.await??;
_=SEM.acquire_many(thread_limit as u32).await.unwrap(); _=SEM.acquire_many(thread_limit as u32).await.unwrap();
Ok(()) Ok(())
} }
async fn cli_roblox_to_snf(paths:Vec<std::path::PathBuf>,output_folder:PathBuf)->AResult<()>{ #[derive(Debug)]
#[allow(dead_code)]
enum ConvertError{
IO(std::io::Error),
SNFMap(strafesnet_snf::map::Error),
RobloxRead(strafesnet_rbx_loader::ReadError),
RobloxLoad(strafesnet_rbx_loader::LoadError),
}
impl std::fmt::Display for ConvertError{
fn fmt(&self,f:&mut std::fmt::Formatter<'_>)->std::fmt::Result{
write!(f,"{self:?}")
}
}
impl std::error::Error for ConvertError{}
async fn convert_to_snf(path:&Path,output_folder:PathBuf)->AResult<()>{
let entire_file=tokio::fs::read(path).await?;
let model=strafesnet_rbx_loader::read(
std::io::Cursor::new(entire_file)
).map_err(ConvertError::RobloxRead)?;
let mut place=model.into_place();
place.run_scripts();
let map=place.to_snf(LoadFailureMode::DefaultToNone).map_err(ConvertError::RobloxLoad)?;
let mut dest=output_folder;
dest.push(path.file_stem().unwrap());
dest.set_extension("snfm");
let file=std::fs::File::create(dest).map_err(ConvertError::IO)?;
strafesnet_snf::map::write_map(file,map).map_err(ConvertError::SNFMap)?;
Ok(())
}
async fn roblox_to_snf(paths:Vec<std::path::PathBuf>,output_folder:PathBuf)->AResult<()>{
let start=std::time::Instant::now(); let start=std::time::Instant::now();
let thread_limit=std::thread::available_parallelism()?.get(); let thread_limit=std::thread::available_parallelism()?.get();
@@ -427,7 +433,7 @@ mod cli{
while let (Ok(permit),Some(path))=(SEM.acquire().await,it.next()){ while let (Ok(permit),Some(path))=(SEM.acquire().await,it.next()){
let output_folder=output_folder.clone(); let output_folder=output_folder.clone();
tokio::spawn(async move{ tokio::spawn(async move{
let result=cli_convert_to_snf(path.as_path(),output_folder).await; let result=convert_to_snf(path.as_path(),output_folder).await;
drop(permit); drop(permit);
match result{ match result{
Ok(())=>(), Ok(())=>(),
@@ -439,19 +445,4 @@ mod cli{
println!("elapsed={:?}", start.elapsed()); println!("elapsed={:?}", start.elapsed());
Ok(()) Ok(())
}
async fn cli_convert_to_snf(path:&Path,output_folder:PathBuf)->AResult<()>{
let entire_file=tokio::fs::read(path).await?;
let snf_data=convert_to_snf(&entire_file)?;
let mut dest=output_folder;
dest.push(path.file_stem().unwrap());
dest.set_extension("snfm");
tokio::fs::write(dest,snf_data).await?;
Ok(())
}
} }
#[cfg(feature="cli")]
pub use cli::Commands;

View File

@@ -1,116 +1,81 @@
use std::io::Cursor; use std::path::{Path,PathBuf};
use clap::{Args,Subcommand};
use anyhow::Result as AResult;
use strafesnet_deferred_loader::deferred_loader::LoadFailureMode; use strafesnet_deferred_loader::deferred_loader::LoadFailureMode;
// === Public library API === #[derive(Subcommand)]
pub enum Commands{
#[derive(Debug,thiserror::Error)] SourceToSNF(SourceToSNFSubcommand),
pub enum ConvertTextureError{ ExtractTextures(ExtractTexturesSubcommand),
#[error("Vtf error {0:?}")] VPKContents(VPKContentsSubcommand),
Vtf(#[from]vtf::Error), BSPContents(BSPContentsSubcommand),
#[error("DDS create error {0:?}")]
DDS(#[from]image_dds::CreateDdsError),
#[error("DDS write error {0:?}")]
DDSWrite(#[from]image_dds::ddsfile::Error),
} }
/// Convert VTF texture bytes to DDS texture bytes. #[derive(Args)]
pub fn convert_texture_to_dds(vtf_data:&[u8])->Result<Vec<u8>,ConvertTextureError>{ pub struct SourceToSNFSubcommand {
let vtf_data=vtf_data.to_vec(); #[arg(long)]
let image=vtf::from_bytes(&vtf_data)?.highres_image.decode(0)?.to_rgba8(); output_folder:PathBuf,
#[arg(required=true)]
let format=if image.width()%4!=0||image.height()%4!=0{ input_files:Vec<PathBuf>,
image_dds::ImageFormat::Rgba8UnormSrgb }
}else{ #[derive(Args)]
image_dds::ImageFormat::BC7RgbaUnormSrgb pub struct ExtractTexturesSubcommand {
}; #[arg(long)]
bsp_file:PathBuf,
let dds=image_dds::dds_from_image( #[arg(long)]
&image, vpk_dir_files:Vec<PathBuf>
format, }
image_dds::Quality::Slow, #[derive(Args)]
image_dds::Mipmaps::GeneratedAutomatic, pub struct VPKContentsSubcommand {
)?; #[arg(long)]
input_file:PathBuf,
let mut buf=Vec::new(); }
dds.write(&mut Cursor::new(&mut buf))?; #[derive(Args)]
Ok(buf) pub struct BSPContentsSubcommand {
#[arg(long)]
input_file:PathBuf,
} }
#[derive(Debug,thiserror::Error)] impl Commands{
pub enum ConvertError{ pub async fn run(self)->AResult<()>{
#[error("BSP read error {0:?}")] match self{
BspRead(strafesnet_bsp_loader::ReadError), Commands::ExtractTextures(subcommand)=>extract_textures(vec![subcommand.bsp_file],subcommand.vpk_dir_files),
#[error("BSP load error {0:?}")] Commands::SourceToSNF(subcommand)=>source_to_snf(subcommand.input_files,subcommand.output_folder).await,
BspLoad(strafesnet_bsp_loader::LoadError), Commands::VPKContents(subcommand)=>vpk_contents(subcommand.input_file),
#[error("SNF map error {0:?}")] Commands::BSPContents(subcommand)=>bsp_contents(subcommand.input_file),
SNFMap(strafesnet_snf::map::Error), }
#[error("BSP parse error {0:?}")] }
BspParse(#[from]vbsp::BspError),
} }
/// Convert a Source BSP file (bytes) to SNF map format (bytes).
pub fn convert_to_snf(bsp_data:&[u8],vpk_list:&[vpk::VPK])->Result<Vec<u8>,ConvertError>{
let bsp=strafesnet_bsp_loader::read(
Cursor::new(bsp_data)
).map_err(ConvertError::BspRead)?;
let map=bsp.to_snf(LoadFailureMode::DefaultToNone,vpk_list).map_err(ConvertError::BspLoad)?; enum VMTContent{
let mut buf=Vec::new();
strafesnet_snf::map::write_map(Cursor::new(&mut buf),map).map_err(ConvertError::SNFMap)?;
Ok(buf)
}
/// Read VPK archives from paths. Useful for loading VPKs needed by `convert_to_snf`.
pub async fn read_vpks(vpk_paths:Vec<std::path::PathBuf>,thread_limit:usize)->Vec<vpk::VPK>{
use futures::StreamExt;
futures::stream::iter(vpk_paths).map(|vpk_path|async{
tokio::task::spawn_blocking(move||vpk::VPK::read(&vpk_path)).await.unwrap().unwrap()
})
.buffer_unordered(thread_limit)
.collect().await
}
// === CLI ===
#[cfg(feature="cli")]
mod cli{
use super::*;
use std::path::{Path,PathBuf};
use std::borrow::Cow;
use clap::{Args,Subcommand};
use anyhow::Result as AResult;
use strafesnet_bsp_loader::loader::BspFinder;
use strafesnet_deferred_loader::loader::Loader;
use strafesnet_deferred_loader::deferred_loader::{MeshDeferredLoader,RenderConfigDeferredLoader};
enum VMTContent{
VMT(String), VMT(String),
VTF(String), VTF(String),
Patch(vmt_parser::material::PatchMaterial), Patch(vmt_parser::material::PatchMaterial),
Unsupported, Unsupported,//don't want to deal with whatever vmt variant
Unresolved, Unresolved,//could not locate a texture because of vmt content
} }
impl VMTContent{ impl VMTContent{
fn vtf(opt:Option<String>)->Self{ fn vtf(opt:Option<String>)->Self{
match opt{ match opt{
Some(s)=>Self::VTF(s), Some(s)=>Self::VTF(s),
None=>Self::Unresolved, None=>Self::Unresolved,
} }
} }
} }
fn get_some_texture(material:vmt_parser::material::Material)->VMTContent{ fn get_some_texture(material:vmt_parser::material::Material)->AResult<VMTContent>{
match material{ //just grab some texture from somewhere for now
Ok(match material{
vmt_parser::material::Material::LightMappedGeneric(mat)=>VMTContent::vtf(Some(mat.base_texture)), vmt_parser::material::Material::LightMappedGeneric(mat)=>VMTContent::vtf(Some(mat.base_texture)),
vmt_parser::material::Material::VertexLitGeneric(mat)=>VMTContent::vtf(mat.base_texture.or(mat.decal_texture)), vmt_parser::material::Material::VertexLitGeneric(mat)=>VMTContent::vtf(mat.base_texture.or(mat.decal_texture)),//this just dies if there is none
vmt_parser::material::Material::VertexLitGenericDx6(mat)=>VMTContent::vtf(mat.base_texture.or(mat.decal_texture)), vmt_parser::material::Material::VertexLitGenericDx6(mat)=>VMTContent::vtf(mat.base_texture.or(mat.decal_texture)),
vmt_parser::material::Material::UnlitGeneric(mat)=>VMTContent::vtf(mat.base_texture), vmt_parser::material::Material::UnlitGeneric(mat)=>VMTContent::vtf(mat.base_texture),
vmt_parser::material::Material::UnlitTwoTexture(mat)=>VMTContent::vtf(mat.base_texture), vmt_parser::material::Material::UnlitTwoTexture(mat)=>VMTContent::vtf(mat.base_texture),
vmt_parser::material::Material::Water(mat)=>VMTContent::vtf(mat.base_texture), vmt_parser::material::Material::Water(mat)=>VMTContent::vtf(mat.base_texture),
vmt_parser::material::Material::WorldVertexTransition(mat)=>VMTContent::vtf(Some(mat.base_texture)), vmt_parser::material::Material::WorldVertexTransition(mat)=>VMTContent::vtf(Some(mat.base_texture)),
vmt_parser::material::Material::EyeRefract(mat)=>VMTContent::vtf(Some(mat.cornea_texture)), vmt_parser::material::Material::EyeRefract(mat)=>VMTContent::vtf(Some(mat.cornea_texture)),
vmt_parser::material::Material::SubRect(mat)=>VMTContent::VMT(mat.material), vmt_parser::material::Material::SubRect(mat)=>VMTContent::VMT(mat.material),//recursive
vmt_parser::material::Material::Sprite(mat)=>VMTContent::vtf(Some(mat.base_texture)), vmt_parser::material::Material::Sprite(mat)=>VMTContent::vtf(Some(mat.base_texture)),
vmt_parser::material::Material::SpriteCard(mat)=>VMTContent::vtf(mat.base_texture), vmt_parser::material::Material::SpriteCard(mat)=>VMTContent::vtf(mat.base_texture),
vmt_parser::material::Material::Cable(mat)=>VMTContent::vtf(Some(mat.base_texture)), vmt_parser::material::Material::Cable(mat)=>VMTContent::vtf(Some(mat.base_texture)),
@@ -120,303 +85,231 @@ mod cli{
vmt_parser::material::Material::Sky(mat)=>VMTContent::vtf(Some(mat.base_texture)), vmt_parser::material::Material::Sky(mat)=>VMTContent::vtf(Some(mat.base_texture)),
vmt_parser::material::Material::Replacements(_mat)=>VMTContent::Unsupported, vmt_parser::material::Material::Replacements(_mat)=>VMTContent::Unsupported,
vmt_parser::material::Material::Patch(mat)=>VMTContent::Patch(mat), vmt_parser::material::Material::Patch(mat)=>VMTContent::Patch(mat),
_=>unreachable!(), _=>return Err(anyhow::Error::msg("vmt failed to parse")),
} })
} }
#[derive(Debug,thiserror::Error)] fn get_vmt<F:Fn(String)->AResult<Option<Vec<u8>>>>(find_stuff:&F,search_name:String)->AResult<vmt_parser::material::Material>{
enum GetVMTError{ if let Some(stuff)=find_stuff(search_name)?{
#[error("Bsp error {0:?}")] //decode vmt and then write
Bsp(#[from]vbsp::BspError), let stuff=String::from_utf8(stuff)?;
#[error("Utf8 error {0:?}")] let material=vmt_parser::from_str(stuff.as_str())?;
Utf8(#[from]std::str::Utf8Error), println!("vmt material={:?}",material);
#[error("Vdf error {0:?}")] return Ok(material);
Vdf(#[from]vmt_parser::VdfError),
#[error("Vmt not found")]
NotFound,
} }
Err(anyhow::Error::msg("vmt not found"))
}
fn get_vmt(finder:BspFinder,search_name:&str)->Result<vmt_parser::material::Material,GetVMTError>{ fn recursive_vmt_loader<F:Fn(String)->AResult<Option<Vec<u8>>>>(find_stuff:&F,material:vmt_parser::material::Material)->AResult<Option<Vec<u8>>>{
let vmt_data=finder.find(search_name)?.ok_or(GetVMTError::NotFound)?; match get_some_texture(material)?{
let vmt_str=core::str::from_utf8(&vmt_data)?; VMTContent::VMT(s)=>recursive_vmt_loader(find_stuff,get_vmt(find_stuff,s)?),
let material=vmt_parser::from_str(vmt_str)?;
Ok(material)
}
#[derive(Debug,thiserror::Error)]
enum LoadVMTError{
#[error("Bsp error {0:?}")]
Bsp(#[from]vbsp::BspError),
#[error("GetVMT error {0:?}")]
GetVMT(#[from]GetVMTError),
#[error("FromUtf8 error {0:?}")]
FromUtf8(#[from]std::string::FromUtf8Error),
#[error("Vdf error {0:?}")]
Vdf(#[from]vmt_parser::VdfError),
#[error("Vmt unsupported")]
Unsupported,
#[error("Vmt unresolved")]
Unresolved,
#[error("Vmt not found")]
NotFound,
}
fn recursive_vmt_loader<'bsp,'vpk,'a>(finder:BspFinder<'bsp,'vpk>,material:vmt_parser::material::Material)->Result<Option<Cow<'a,[u8]>>,LoadVMTError>
where
'bsp:'a,
'vpk:'a,
{
match get_some_texture(material){
VMTContent::VMT(s)=>recursive_vmt_loader(finder,get_vmt(finder,s.as_str())?),
VMTContent::VTF(s)=>{ VMTContent::VTF(s)=>{
let mut texture_file_name=PathBuf::from("materials"); let mut texture_file_name=PathBuf::from("materials");
texture_file_name.push(s); texture_file_name.push(s);
texture_file_name.set_extension("vtf"); texture_file_name.set_extension("vtf");
Ok(finder.find(texture_file_name.to_str().unwrap())?) find_stuff(texture_file_name.into_os_string().into_string().unwrap())
}, },
VMTContent::Patch(mat)=>recursive_vmt_loader(finder, VMTContent::Patch(mat)=>recursive_vmt_loader(find_stuff,
mat.resolve(|search_name| mat.resolve(|search_name|{
match finder.find(search_name)?{ match find_stuff(search_name.to_string())?{
Some(bytes)=>Ok(String::from_utf8(bytes.into_owned())?), Some(bytes)=>Ok(String::from_utf8(bytes)?),
None=>Err(LoadVMTError::NotFound), None=>Err(anyhow::Error::msg("could not find vmt")),
} }
)? })?
), ),
VMTContent::Unsupported=>Err(LoadVMTError::Unsupported), VMTContent::Unsupported=>{println!("Unsupported vmt");Ok(None)},//print and move on
VMTContent::Unresolved=>Err(LoadVMTError::Unresolved), VMTContent::Unresolved=>{println!("Unresolved vmt");Ok(None)},
}
}
fn load_texture<'bsp,'vpk,'a>(finder:BspFinder<'bsp,'vpk>,texture_name:&str)->Result<Option<Cow<'a,[u8]>>,LoadVMTError>
where
'bsp:'a,
'vpk:'a,
{
let mut texture_file_name=PathBuf::from("materials");
let texture_file_name_lowercase=texture_name.to_lowercase();
texture_file_name.push(texture_file_name_lowercase.clone());
let stem=PathBuf::from(texture_file_name.file_stem().unwrap());
texture_file_name.pop();
texture_file_name.push(stem);
if let Some(stuff)=finder.find(texture_file_name.to_str().unwrap())?{
return Ok(Some(stuff))
}
let mut texture_file_name_vmt=texture_file_name.clone();
texture_file_name.set_extension("vtf");
texture_file_name_vmt.set_extension("vmt");
recursive_vmt_loader(finder,get_vmt(finder,texture_file_name_vmt.to_str().unwrap())?)
} }
}
#[derive(Subcommand)] fn extract_textures(paths:Vec<PathBuf>,vpk_paths:Vec<PathBuf>)->AResult<()>{
pub enum Commands{ std::fs::create_dir_all("textures")?;
SourceToSNF(SourceToSNFSubcommand), let vpk_list:Vec<vpk::VPK>=vpk_paths.into_iter().map(|vpk_path|vpk::VPK::read(&vpk_path).expect("vpk file does not exist")).collect();
ExtractTextures(ExtractTexturesSubcommand), for path in paths{
VPKContents(VPKContentsSubcommand), let mut deduplicate=std::collections::HashSet::new();
BSPContents(BSPContentsSubcommand), let bsp=vbsp::Bsp::read(std::fs::read(path)?.as_ref())?;
}
#[derive(Args)]
pub struct SourceToSNFSubcommand {
#[arg(long)]
output_folder:PathBuf,
#[arg(required=true)]
input_files:Vec<PathBuf>,
#[arg(long)]
vpks:Vec<PathBuf>,
}
#[derive(Args)]
pub struct ExtractTexturesSubcommand{
#[arg(required=true)]
bsp_files:Vec<PathBuf>,
#[arg(long)]
vpks:Vec<PathBuf>,
}
#[derive(Args)]
pub struct VPKContentsSubcommand {
#[arg(long)]
input_file:PathBuf,
}
#[derive(Args)]
pub struct BSPContentsSubcommand {
#[arg(long)]
input_file:PathBuf,
}
impl Commands{
pub async fn run(self)->AResult<()>{
match self{
Commands::SourceToSNF(subcommand)=>cli_source_to_snf(subcommand.input_files,subcommand.output_folder,subcommand.vpks).await,
Commands::ExtractTextures(subcommand)=>cli_extract_textures(subcommand.bsp_files,subcommand.vpks).await,
Commands::VPKContents(subcommand)=>vpk_contents(subcommand.input_file),
Commands::BSPContents(subcommand)=>bsp_contents(subcommand.input_file),
}
}
}
#[derive(Debug,thiserror::Error)]
enum ExtractTextureError{
#[error("Io error {0:?}")]
Io(#[from]std::io::Error),
#[error("Bsp error {0:?}")]
Bsp(#[from]vbsp::BspError),
#[error("MeshLoad error {0:?}")]
MeshLoad(#[from]strafesnet_bsp_loader::loader::MeshError),
#[error("Load VMT error {0:?}")]
LoadVMT(#[from]LoadVMTError),
}
async fn gimme_them_textures(path:&Path,vpk_list:&[vpk::VPK],send_texture:tokio::sync::mpsc::Sender<(Vec<u8>,String)>)->Result<(),ExtractTextureError>{
let bsp=vbsp::Bsp::read(tokio::fs::read(path).await?.as_ref())?;
let loader_bsp=strafesnet_bsp_loader::Bsp::new(bsp);
let bsp=loader_bsp.as_ref();
let mut texture_deferred_loader=RenderConfigDeferredLoader::new();
for texture in bsp.textures(){ for texture in bsp.textures(){
texture_deferred_loader.acquire_render_config_id(Some(Cow::Borrowed(texture.name()))); deduplicate.insert(PathBuf::from(texture.name()));
} }
//dedupe prop models
let mut mesh_deferred_loader=MeshDeferredLoader::new(); let mut model_dedupe=std::collections::HashSet::new();
for prop in bsp.static_props(){ for prop in bsp.static_props(){
mesh_deferred_loader.acquire_mesh_id(prop.model()); model_dedupe.insert(prop.model());
} }
let finder=BspFinder{ //grab texture names from props
bsp:&loader_bsp, for model_name in model_dedupe{
vpks:vpk_list //.mdl, .vvd, .dx90.vtx
}; let mut path=PathBuf::from(model_name);
let file_name=PathBuf::from(path.file_stem().unwrap());
let mut mesh_loader=strafesnet_bsp_loader::loader::ModelLoader::new(finder); path.pop();
for model_path in mesh_deferred_loader.into_indices(){ path.push(file_name);
let model:vmdl::Model=match mesh_loader.load(model_path){ let mut vvd_path=path.clone();
Ok(model)=>model, let mut vtx_path=path.clone();
Err(e)=>{ vvd_path.set_extension("vvd");
println!("Model={model_path} Load model error: {e}"); vtx_path.set_extension("dx90.vtx");
continue; match (bsp.pack.get(model_name),bsp.pack.get(vvd_path.as_os_str().to_str().unwrap()),bsp.pack.get(vtx_path.as_os_str().to_str().unwrap())){
}, (Ok(Some(mdl_file)),Ok(Some(vvd_file)),Ok(Some(vtx_file)))=>{
}; match (vmdl::mdl::Mdl::read(mdl_file.as_ref()),vmdl::vvd::Vvd::read(vvd_file.as_ref()),vmdl::vtx::Vtx::read(vtx_file.as_ref())){
(Ok(mdl),Ok(vvd),Ok(vtx))=>{
let model=vmdl::Model::from_parts(mdl,vtx,vvd);
for texture in model.textures(){ for texture in model.textures(){
for search_path in &texture.search_paths{ for search_path in &texture.search_paths{
let mut path=PathBuf::from(search_path.as_str()); let mut path=PathBuf::from(search_path.as_str());
path.push(texture.name.as_str()); path.push(texture.name.as_str());
let path=path.to_str().unwrap().to_owned(); deduplicate.insert(path);
texture_deferred_loader.acquire_render_config_id(Some(Cow::Owned(path)));
} }
} }
},
_=>println!("model_name={} error",model_name),
} }
},
for texture_path in texture_deferred_loader.into_indices(){ _=>println!("no model name={}",model_name),
match load_texture(finder,&texture_path){
Ok(Some(texture))=>send_texture.send(
(texture.into_owned(),texture_path.into_owned())
).await.unwrap(),
Ok(None)=>(),
Err(e)=>println!("Texture={texture_path} Load error: {e}"),
} }
} }
Ok(()) let pack=&bsp.pack;
} let vpk_list=&vpk_list;
std::thread::scope(move|s|{
let mut thread_handles=Vec::new();
for texture_name in deduplicate{
let mut found_texture=false;
//LMAO imagine having to write type names
let write_image=|mut stuff,write_file_name|{
let image=vtf::from_bytes(&mut stuff)?.highres_image.decode(0)?.to_rgba8();
#[derive(Debug,thiserror::Error)] let format=if image.width()%4!=0||image.height()%4!=0{
enum CliConvertTextureError{ image_dds::ImageFormat::Rgba8UnormSrgb
#[error("IO error {0:?}")] }else{
Io(#[from]std::io::Error), image_dds::ImageFormat::BC7RgbaUnormSrgb
#[error("Convert texture error {0:?}")] };
Convert(#[from]ConvertTextureError), //this fails if the image dimensions are not a multiple of 4
} let dds = image_dds::dds_from_image(
&image,
async fn cli_convert_texture(texture:Vec<u8>,write_file_name:impl AsRef<Path>)->Result<(),CliConvertTextureError>{ format,
let dds_data=convert_texture_to_dds(&texture)?; image_dds::Quality::Slow,
image_dds::Mipmaps::GeneratedAutomatic,
)?;
//write dds
let mut dest=PathBuf::from("textures"); let mut dest=PathBuf::from("textures");
dest.push(write_file_name); dest.push(write_file_name);
dest.set_extension("dds"); dest.set_extension("dds");
std::fs::create_dir_all(dest.parent().unwrap())?; std::fs::create_dir_all(dest.parent().unwrap())?;
let mut writer=std::io::BufWriter::new(std::fs::File::create(dest)?); let mut writer = std::io::BufWriter::new(std::fs::File::create(dest)?);
std::io::Write::write_all(&mut writer,&dds_data)?; dds.write(&mut writer)?;
Ok::<(),anyhow::Error>(())
};
let find_stuff=|search_file_name:String|{
println!("search_file_name={}",search_file_name);
match pack.get(search_file_name.as_str())?{
Some(file)=>return Ok(Some(file)),
_=>(),
}
//search pak list
for vpk_index in vpk_list{
if let Some(vpk_entry)=vpk_index.tree.get(search_file_name.as_str()){
return Ok(Some(match vpk_entry.get()?{
std::borrow::Cow::Borrowed(bytes)=>bytes.to_vec(),
std::borrow::Cow::Owned(bytes)=>bytes,
}));
}
}
Ok::<Option<Vec<u8>>,anyhow::Error>(None)
};
let loader=|texture_name:String|{
let mut texture_file_name=PathBuf::from("materials");
//lower case
let texture_file_name_lowercase=texture_name.to_lowercase();
texture_file_name.push(texture_file_name_lowercase.clone());
//remove stem and search for both vtf and vmt files
let stem=PathBuf::from(texture_file_name.file_stem().unwrap());
texture_file_name.pop();
texture_file_name.push(stem);
//somehow search for both files
let mut texture_file_name_vmt=texture_file_name.clone();
texture_file_name.set_extension("vtf");
texture_file_name_vmt.set_extension("vmt");
if let Some(stuff)=find_stuff(texture_file_name.to_string_lossy().to_string())?{
return Ok(Some(stuff))
}
recursive_vmt_loader(&find_stuff,get_vmt(&find_stuff,texture_file_name_vmt.to_string_lossy().to_string())?)
};
if let Some(stuff)=loader(texture_name.to_string_lossy().to_string())?{
found_texture=true;
let texture_name=texture_name.clone();
thread_handles.push(s.spawn(move||write_image(stuff,texture_name)));
}
if !found_texture{
println!("no data");
}
}
for thread in thread_handles{
match thread.join(){
Ok(Err(e))=>println!("write error: {:?}",e),
Err(e)=>println!("thread error: {:?}",e),
Ok(_)=>(),
}
}
Ok::<(),anyhow::Error>(())
})?
}
Ok(()) Ok(())
} }
async fn cli_extract_textures(paths:Vec<PathBuf>,vpk_paths:Vec<PathBuf>)->AResult<()>{ fn vpk_contents(vpk_path:PathBuf)->AResult<()>{
tokio::try_join!(
tokio::fs::create_dir_all("extracted_textures"),
tokio::fs::create_dir_all("textures"),
tokio::fs::create_dir_all("meshes"),
)?;
let thread_limit=std::thread::available_parallelism()?.get();
let vpk_list=read_vpks(vpk_paths,thread_limit).await;
let vpk_list:&[vpk::VPK]=vpk_list.leak();
let (send_texture,mut recv_texture)=tokio::sync::mpsc::channel(thread_limit);
let mut it=paths.into_iter();
let extract_thread=tokio::spawn(async move{
static SEM:tokio::sync::Semaphore=tokio::sync::Semaphore::const_new(0);
SEM.add_permits(thread_limit);
while let (Ok(permit),Some(path))=(SEM.acquire().await,it.next()){
let send=send_texture.clone();
tokio::spawn(async move{
let result=gimme_them_textures(&path,vpk_list,send).await;
drop(permit);
match result{
Ok(())=>(),
Err(e)=>println!("Map={path:?} Decode error: {e:?}"),
}
});
}
});
static SEM:tokio::sync::Semaphore=tokio::sync::Semaphore::const_new(0);
SEM.add_permits(thread_limit);
while let (Ok(permit),Some((data,dest)))=(SEM.acquire().await,recv_texture.recv().await){
tokio::spawn(async move{
let result=cli_convert_texture(data,dest).await;
drop(permit);
match result{
Ok(())=>(),
Err(e)=>println!("Convert error: {e:?}"),
}
});
}
extract_thread.await?;
_=SEM.acquire_many(thread_limit as u32).await?;
Ok(())
}
fn vpk_contents(vpk_path:PathBuf)->AResult<()>{
let vpk_index=vpk::VPK::read(&vpk_path)?; let vpk_index=vpk::VPK::read(&vpk_path)?;
for (label,entry) in vpk_index.tree.into_iter(){ for (label,entry) in vpk_index.tree.into_iter(){
println!("vpk label={} entry={:?}",label,entry); println!("vpk label={} entry={:?}",label,entry);
} }
Ok(()) Ok(())
} }
fn bsp_contents(path:PathBuf)->AResult<()>{ fn bsp_contents(path:PathBuf)->AResult<()>{
let bsp=vbsp::Bsp::read(std::fs::read(path)?.as_ref())?; let bsp=vbsp::Bsp::read(std::fs::read(path)?.as_ref())?;
for file_name in bsp.pack.into_zip().into_inner().unwrap().file_names(){ for file_name in bsp.pack.into_zip().into_inner().unwrap().file_names(){
println!("file_name={:?}",file_name); println!("file_name={:?}",file_name);
} }
Ok(()) Ok(())
} }
async fn cli_convert_to_snf(path:&Path,vpk_list:&[vpk::VPK],output_folder:PathBuf)->AResult<()>{ #[derive(Debug)]
#[allow(dead_code)]
enum ConvertError{
IO(std::io::Error),
SNFMap(strafesnet_snf::map::Error),
BspRead(strafesnet_bsp_loader::ReadError),
BspLoad(strafesnet_bsp_loader::LoadError),
}
impl std::fmt::Display for ConvertError{
fn fmt(&self,f:&mut std::fmt::Formatter<'_>)->std::fmt::Result{
write!(f,"{self:?}")
}
}
impl std::error::Error for ConvertError{}
async fn convert_to_snf(path:&Path,output_folder:PathBuf)->AResult<()>{
let entire_file=tokio::fs::read(path).await?; let entire_file=tokio::fs::read(path).await?;
let snf_data=convert_to_snf(&entire_file,vpk_list)?;
let bsp=strafesnet_bsp_loader::read(
std::io::Cursor::new(entire_file)
).map_err(ConvertError::BspRead)?;
let map=bsp.to_snf(LoadFailureMode::DefaultToNone).map_err(ConvertError::BspLoad)?;
let mut dest=output_folder; let mut dest=output_folder;
dest.push(path.file_stem().unwrap()); dest.push(path.file_stem().unwrap());
dest.set_extension("snfm"); dest.set_extension("snfm");
tokio::fs::write(dest,snf_data).await?; let file=std::fs::File::create(dest).map_err(ConvertError::IO)?;
strafesnet_snf::map::write_map(file,map).map_err(ConvertError::SNFMap)?;
Ok(()) Ok(())
} }
async fn source_to_snf(paths:Vec<std::path::PathBuf>,output_folder:PathBuf)->AResult<()>{
async fn cli_source_to_snf(paths:Vec<std::path::PathBuf>,output_folder:PathBuf,vpk_paths:Vec<PathBuf>)->AResult<()>{
let start=std::time::Instant::now(); let start=std::time::Instant::now();
let thread_limit=std::thread::available_parallelism()?.get(); let thread_limit=std::thread::available_parallelism()?.get();
let vpk_list=read_vpks(vpk_paths,thread_limit).await;
let vpk_list:&[vpk::VPK]=vpk_list.leak();
let mut it=paths.into_iter(); let mut it=paths.into_iter();
static SEM:tokio::sync::Semaphore=tokio::sync::Semaphore::const_new(0); static SEM:tokio::sync::Semaphore=tokio::sync::Semaphore::const_new(0);
SEM.add_permits(thread_limit); SEM.add_permits(thread_limit);
@@ -424,7 +317,7 @@ mod cli{
while let (Ok(permit),Some(path))=(SEM.acquire().await,it.next()){ while let (Ok(permit),Some(path))=(SEM.acquire().await,it.next()){
let output_folder=output_folder.clone(); let output_folder=output_folder.clone();
tokio::spawn(async move{ tokio::spawn(async move{
let result=cli_convert_to_snf(path.as_path(),vpk_list,output_folder).await; let result=convert_to_snf(path.as_path(),output_folder).await;
drop(permit); drop(permit);
match result{ match result{
Ok(())=>(), Ok(())=>(),
@@ -436,7 +329,4 @@ mod cli{
println!("elapsed={:?}", start.elapsed()); println!("elapsed={:?}", start.elapsed());
Ok(()) Ok(())
}
} }
#[cfg(feature="cli")]
pub use cli::Commands;