userDaniel Pcancel
reporust_rebootcancel

1,623 Commits over 396 Days - 0.17cph!

4 Months Ago
Update: ServerProfiler - initial FixedStorage support - Added test for FixedStorage snapshot recording - update ProfileExporter to handle FixedStorage binary stream quirks - updated binaries based on 2ce19cfe New mode will allow us to stream profiling info until the buffer fills up. RCon commands will come next Tests: unit tests
4 Months Ago
Merge: from profiling_improvements - Buildfix for tests trying to use ServerProfiler in non-server env Tests: scripts compile in editor CLIENT mode
4 Months Ago
Buildfix: exclude ServerProfiler tests from non-server builds Tests: built CLIENT config in editor
4 Months Ago
Merge: from profiling_improvements - Rewrote internal storage of profiling data to use 1 buffer per thread - Bugfix for allocation graphs not properly resetting and having gaps at the start - Bugfix for failing to generate export in rare cases where mono runtime allocates with no managed callstack Tests: unit tests and generating snapshots in editor
4 Months Ago
Bugfix: ProfileExporter.JSON - don't spam allocs-0 counter for worker threads Tests: snapshot in editor on craggy
4 Months Ago
Merge: from main
4 Months Ago
Update: ServerProfiler.Core binaries - Relase bins built from 66537fcc Didn't remember if I snuck in debug bins at one point, so updating them to be safe Tests: unit tests
4 Months Ago
Bugfix: ProfileExporter.JSON - reset allocation graphs - Reset when a new frame starts - Reset on worker threads if it allocates Tests: snapshot in editor on craggy
4 Months Ago
Clean: ProfileExporter.JSON - don't cache per-frame callstack depths Was never used, so don't waste allocs. Tests: none, trivial change
4 Months Ago
Clean: ProfileExporter.JSON - remove debug logs Tests: none, trivial change
4 Months Ago
Bugfix: ProfileExporter.JSON - gracefully handle managed allocations coming from native runtime - Emit "<mono-native-runtime>" if we don't have managed callstack Finally caught it - this can happen when mono tries to invoke a managed callback which requires a managed allocation (the callback accepts string[], for example) as a first method in managed code. Was able to repro in editor due to it's script compilation callbacks. Tests: triggered perfsnapshot 40 times without issues
4 Months Ago
Bugfix: ProfileExporter - avoid reading allocs at the start of the frame as method-entries - Added a bunch of temporary logging to help track down last issue Rare, but legal due to our filtering of code. Tests: snapshotted a bunch of times in editor (there's still one issue with main thread export)
4 Months Ago
Update: ProfileBinViewer displays binary offsets for marks Tests: opened a couple bin snapshots
4 Months Ago
Bugfix: avoid out-of-bounds access when scanning for alloc-only threads Fixes perf snapshot export failing to generate while processing worker threads. There's still a rare case of main thread crashing - investigating Tests: Exported a snapshot in editor
5 Months Ago
Update: ProfileBinViewer can now grok the new profile .bin files Tests: opened bin profile from editor
5 Months Ago
Update: rewrite ServerProfile exporters to work with new data layour - expanded unit test - Core binary should be the same, but copied from the freshest release build - purposefully borked ProfileBinViewer - will fix next Tests: unit tests & took a snapshot on craggy
5 Months Ago
Update: Rewrote how ServerProfiler.Core stores profiling data - got a 2% improvement on 1mil empty call benchmark It will be easier to add new features to the profiler (user scopes, user counters, continuous capture). Need to rewrite the exporters now. Tests: unit tests
5 Months Ago
Tests: perf test to measure ServerProfiler overhead Going to try a different internal storage approach to make future changes easier (and potentially faster) Tests: ran the new tests
5 Months Ago
Merge: from parallel_validatemove - Removes PlayerCache.ValidPlayers allocs Tests: took a snapshot on Craggy in editor
5 Months Ago
Optim: PlayerCache.ValidPlayers no longer allocates garbage Tests: took snapshot on Craggy in editor to confirm
5 Months Ago
Merge: from main Tests: none, no conflicts
5 Months Ago
Merge: from profiling_improvements - Reduces overhead of serializing to/from ProtoBuf by not tracking BufferStream calls Tests: snapshot on Craggy in editor
5 Months Ago
Update: further filtering of methods - Dropping BaseEntity.Is* methods that are just HasFlag wrappers - Dropping new Rust.Data BufferStream and RangeHandle Should reduce overhead on serialization Tests: snapshot on Craggy in editor
5 Months Ago
Merge: from main Tests: none, no conflicts
5 Months Ago
Merge: from relationshipmanager_leaks - Server and Client-side bugfixes for pooling around RelationshipManager types Tests: local 2 player session with flicking open-closed the contacts screen and with pooling tracking
5 Months Ago
Merge: from main Tests: none, no conflicts
5 Months Ago
Bugfix: return ProtoBuf.PlayerRelationships back to pool after client rpc is executed Think it was the only PlayerRelationships leak on client Tests: local session on craggy, opened&closed contacts - saw pool metrics stay stable
5 Months Ago
Update: using new RelationshipManager.ClearRelations to avoid potential pooling leaks Tests: none
5 Months Ago
Bugfix: clear PlayerRelationshipInfo when returning PlayerRelationship to pool Tests: none, since it turns out that we don't have any code that stresses this path - one more thing to fix then
5 Months Ago
Bugfix: return PlayerRelationshipInfo back to pool when forgetting a player 2 more places to fix Tests: none
5 Months Ago
Bugfix: return to pool ProtoBuf.RelationshipManager.PlayerRelationships + nested types after client rpcs There's still one more server leak and a client leak Tests: local 2-player session. Made sure changed code is being stepped through.
5 Months Ago
Merge: from main Tests: none, no conflicts
5 Months Ago
Merge: from reduce_appmarkersellorder_allocs - Reduces pool spillage and misses of ProtoBuf.AppMarker.SellOrder Tests: started server on procgen map
5 Months Ago
Optim: Increase ProtoBuf.AppMarker.SellOrder pool capacity to 2k - Should reduce creation and spillage of SellOrders Tests: booted server on procgen.
5 Months Ago
Merge: from profiling_improvements Tests: snapshot on craggy in editor
5 Months Ago
Merge: from main Tests: none, no conflicts
5 Months Ago
Update: Further removal of about 10% methods from perf snapshot - Built using e70c083b Tests: took a snapshot on craggy
5 Months Ago
Merge: from listcompare_optim - ListHashSet can now be pooled and supports Compare - Optim Networkable.UpdateSubscriptions via ListHashSet Tests: unit tests
5 Months Ago
Merge: from main Tests: none, no conflicts
5 Months Ago
Clean: simplify code via ?. notation Tests: unit tests
5 Months Ago
Optim: Constain List<T>.Compare to only works on List<T> This removes remaining 4 boxing allocations Tests: unit tests
5 Months Ago
Optim: Use ListHashSet<T>.Compare in UpdateSubscriptions and UpdateHighPrioritySubscriptions Tests: none, trivial change
5 Months Ago
Update: Network.Visibility.Provider works on ListHashSet Tests: none, simple changes
5 Months Ago
Update: Facepunch.Pool now supports ListHashSet - gave ListHashSet a default ctor Tests: unit tests
5 Months Ago
Update: ListHashSet has it's own specialized static Compare method - Replciated tests from ListExtensionTests for ListHashSet Tests: ran unit tests
5 Months Ago
Optim: List.Compare now uses pooled hashsets instead of hashing-and-sorting 2x faster. Now, just need to see if I can get rid of these remaining allocs (just 4 per stable run) Tests: unit tests
5 Months Ago
Merge: from profiling_improvements - Reduce capture size by ~19% by filtering out more methods Tests: snapshot on craggy in editor
5 Months Ago
Update: further reduce capture scope by ~19% - Built from c4679e41 Tests: snapshot in editor on Craggy
5 Months Ago
Merge: from main
5 Months Ago
Merge: from parallel_validatemove - Bugfix: GamePhysics.OverlapCapsules no longer skips 0-length-capsule queries - Additional unit tests Tests: unit tests + staging demo playback