// Copyright Epic Games, Inc. All Rights Reserved. #include "DerivedDataBackendInterface.h" #if WITH_HTTP_DDC_BACKEND #include "Algo/Accumulate.h" #include "Algo/Find.h" #include "Algo/Transform.h" #include "Compression/CompressedBuffer.h" #include "Containers/DepletableMpscQueue.h" #include "Containers/StaticArray.h" #include "Containers/StringView.h" #include "Containers/Ticker.h" #include "DerivedDataCacheKey.h" #include "DerivedDataCachePrivate.h" #include "DerivedDataCacheRecord.h" #include "DerivedDataCacheUsageStats.h" #include "DerivedDataChunk.h" #include "DerivedDataRequest.h" #include "DerivedDataRequestOwner.h" #include "DerivedDataValue.h" #include "Dom/JsonObject.h" #include "Experimental/Async/LazyEvent.h" #include "Experimental/Containers/FAAArrayQueue.h" #include "GenericPlatform/GenericPlatformFile.h" #include "HAL/IConsoleManager.h" #include "HAL/PlatformFileManager.h" #include "HAL/Thread.h" #include "IO/IoHash.h" #include "Memory/SharedBuffer.h" #include "Misc/CString.h" #include "Misc/FileHelper.h" #include "Misc/Optional.h" #include "Misc/ScopeLock.h" #include "Misc/SecureHash.h" #include "Misc/StringBuilder.h" #include "Policies/CondensedJsonPrintPolicy.h" #include "ProfilingDebugging/CountersTrace.h" #include "ProfilingDebugging/CpuProfilerTrace.h" #include "Serialization/BufferArchive.h" #include "Serialization/CompactBinary.h" #include "Serialization/CompactBinaryPackage.h" #include "Serialization/CompactBinaryValidation.h" #include "Serialization/CompactBinaryWriter.h" #include "Serialization/JsonReader.h" #include "Serialization/JsonSerializer.h" #include "Serialization/JsonWriter.h" #include "Tasks/Task.h" #include "Http/HttpClient.h" #if PLATFORM_MICROSOFT #include "Microsoft/WindowsHWrapper.h" #include "Microsoft/AllowMicrosoftPlatformTypes.h" #include #include #include "Microsoft/HideMicrosoftPlatformTypes.h" #else #include #include #include #endif // Enables data request helpers that internally // batch requests to reduce the number of concurrent // connections. #ifndef WITH_DATAREQUEST_HELPER #define WITH_DATAREQUEST_HELPER 1 #endif #define UE_HTTPDDC_GET_REQUEST_POOL_SIZE 48 #define UE_HTTPDDC_PUT_REQUEST_POOL_SIZE 16 #define UE_HTTPDDC_NONBLOCKING_REQUEST_POOL_SIZE 128 #define UE_HTTPDDC_MAX_FAILED_LOGIN_ATTEMPTS 16 #define UE_HTTPDDC_MAX_ATTEMPTS 4 #define UE_HTTPDDC_BATCH_SIZE 12 #define UE_HTTPDDC_BATCH_NUM 64 #define UE_HTTPDDC_BATCH_GET_WEIGHT 4 #define UE_HTTPDDC_BATCH_HEAD_WEIGHT 1 #define UE_HTTPDDC_BATCH_WEIGHT_HINT 12 namespace UE::DerivedData { TRACE_DECLARE_INT_COUNTER(HttpDDC_Exist, TEXT("HttpDDC Exist")); TRACE_DECLARE_INT_COUNTER(HttpDDC_ExistHit, TEXT("HttpDDC Exist Hit")); TRACE_DECLARE_INT_COUNTER(HttpDDC_Get, TEXT("HttpDDC Get")); TRACE_DECLARE_INT_COUNTER(HttpDDC_GetHit, TEXT("HttpDDC Get Hit")); TRACE_DECLARE_INT_COUNTER(HttpDDC_Put, TEXT("HttpDDC Put")); TRACE_DECLARE_INT_COUNTER(HttpDDC_PutHit, TEXT("HttpDDC Put Hit")); TRACE_DECLARE_INT_COUNTER(HttpDDC_BytesReceived, TEXT("HttpDDC Bytes Received")); TRACE_DECLARE_INT_COUNTER(HttpDDC_BytesSent, TEXT("HttpDDC Bytes Sent")); template class TRefCountedUniqueFunction final : public FThreadSafeRefCountedObject { public: explicit TRefCountedUniqueFunction(T&& InFunction) : Function(MoveTemp(InFunction)) { } const T& GetFunction() const { return Function; } private: T Function; }; //---------------------------------------------------------------------------------------------------------- // Forward declarations //---------------------------------------------------------------------------------------------------------- bool VerifyPayload(const FSHAHash& Hash, const TCHAR* Namespace, const TCHAR* Bucket, const TCHAR* CacheKey, const TArray64& Payload); bool VerifyPayload(const FIoHash& Hash, const TCHAR* Namespace, const TCHAR* Bucket, const TCHAR* CacheKey, const TArray64& Payload); bool VerifyRequest(const class FHttpRequest* Request, const TCHAR* Namespace, const TCHAR* Bucket, const TCHAR* CacheKey, const TArray64& Payload); bool HashPayload(class FHttpRequest* Request, const TArrayView Payload); bool ShouldAbortForShutdown(); #if WITH_DATAREQUEST_HELPER //---------------------------------------------------------------------------------------------------------- // FDataRequestHelper //---------------------------------------------------------------------------------------------------------- /** * Helper class for requesting data. Will batch requests once the number of concurrent requests reach a threshold. */ struct FDataRequestHelper { FDataRequestHelper(FHttpRequestPool* InPool, const TCHAR* InNamespace, const TCHAR* InBucket, const TCHAR* InCacheKey, TArray64* OutData) : Request(nullptr) , Pool(InPool) , bVerified(false, 1) { Request = Pool->GetFreeRequest(); if (Request && OutData != nullptr) { // We are below the threshold, make the connection immediately. OutData is set so this is a get. FString Uri = FString::Printf(TEXT("api/v1/c/ddc/%s/%s/%s.raw"), InNamespace, InBucket, InCacheKey); const FHttpRequest::EResult Result = Request->PerformBlockingDownload(*Uri, OutData); if (FHttpRequest::IsSuccessResponse(Request->GetResponseCode())) { if (VerifyRequest(Request, InNamespace, InBucket, InCacheKey, *OutData)) { TRACE_COUNTER_ADD(HttpDDC_GetHit, int64(1)); TRACE_COUNTER_ADD(HttpDDC_BytesReceived, int64(Request->GetBytesReceived())); bVerified[0] = true; } } } else if (Request) { // We are below the threshold, make the connection immediately. OutData is missing so this is a head. FString Uri = FString::Printf(TEXT("api/v1/c/ddc/%s/%s/%s"), InNamespace, InBucket, InCacheKey); const FHttpRequest::EResult Result = Request->PerformBlockingHead(*Uri); if (FHttpRequest::IsSuccessResponse(Request->GetResponseCode())) { TRACE_COUNTER_ADD(HttpDDC_ExistHit, int64(1)); bVerified[0] = true; } } else { // We have exceeded the threshold for concurrent connections, start or add this request // to a batched request. if (IsQueueCandidate(1, OutData && !OutData->IsEmpty())) { Request = QueueBatchRequest( InPool, InNamespace, InBucket, TConstArrayView({InCacheKey}), OutData ? TConstArrayView*>({OutData}) : TConstArrayView*>(), bVerified ); } if (!Request) { Request = Pool->WaitForFreeRequest(); FQueuedBatchEntry Entry{ InNamespace, InBucket, TConstArrayView({InCacheKey}), OutData ? TConstArrayView*>({OutData}) : TConstArrayView*>(), OutData && !OutData->IsEmpty() ? FHttpRequest::ERequestVerb::Get : FHttpRequest::ERequestVerb::Head, &bVerified }; PerformBatchQuery(Request, TArrayView(&Entry, 1)); } } } // Constructor specifically for batched head queries FDataRequestHelper(FHttpRequestPool* InPool, const TCHAR* InNamespace, const TCHAR* InBucket, TConstArrayView InCacheKeys) : Request(nullptr) , Pool(InPool) , bVerified(false, InCacheKeys.Num()) { // Transform the FString array to char pointers TArray CacheKeys; Algo::Transform(InCacheKeys, CacheKeys, [](const FString& Key) { return *Key; }); Request = Pool->GetFreeRequest(); if (Request || !IsQueueCandidate(InCacheKeys.Num(), false)) { // If the request is too big for existing batches, wait for a free connection and create our own. if (!Request) { Request = Pool->WaitForFreeRequest(); } FQueuedBatchEntry Entry{ InNamespace, InBucket, CacheKeys, TConstArrayView*>(), FHttpRequest::ERequestVerb::Head, &bVerified }; PerformBatchQuery(Request, TArrayView(&Entry, 1)); } else { Request = QueueBatchRequest( InPool, InNamespace, InBucket, CacheKeys, TConstArrayView*>(), bVerified ); if (!Request) { Request = Pool->WaitForFreeRequest(); FQueuedBatchEntry Entry{ InNamespace, InBucket, CacheKeys, TConstArrayView*>(), FHttpRequest::ERequestVerb::Head, &bVerified }; PerformBatchQuery(Request, TArrayView(&Entry, 1)); } } } ~FDataRequestHelper() { if (Request) { Pool->ReleaseRequestToPool(Request); } } static void StaticInitialize() { static bool bInitialized = false; check(!bInitialized); for (FBatch& Batch : Batches) { Batch.Reserved = 0; Batch.Ready = 0; Batch.Complete = TUniquePtr(FPlatformProcess::GetSynchEventFromPool(true)); } bInitialized = true; } static void StaticShutdown() { for (FBatch& Batch : Batches) { Batch.Complete.Reset(); } } bool IsSuccess() const { return bVerified[0]; } const TBitArray<>& IsBatchSuccess() const { return bVerified; } int64 GetResponseCode() const { return Request ? Request->GetResponseCode() : 0; } private: struct FQueuedBatchEntry { const TCHAR* Namespace; const TCHAR* Bucket; TConstArrayView CacheKeys; TConstArrayView*> OutDatas; FHttpRequest::ERequestVerb Verb; TBitArray<>* bSuccess; }; struct FBatch { struct FEventDeleter { void operator()(FEvent* Event) { FPlatformProcess::ReturnSynchEventToPool(Event); } }; FQueuedBatchEntry Entries[UE_HTTPDDC_BATCH_SIZE]; std::atomic Reserved; std::atomic Ready; std::atomic WeightHint; FHttpRequest* Request; TUniquePtr Complete; }; FHttpRequest* Request; FHttpRequestPool* Pool; TBitArray<> bVerified; static std::atomic FirstAvailableBatch; static TStaticArray Batches; static uint32 ComputeWeight(int32 NumKeys, bool bHasDatas) { return NumKeys * (bHasDatas ? UE_HTTPDDC_BATCH_GET_WEIGHT : UE_HTTPDDC_BATCH_HEAD_WEIGHT); } static bool IsQueueCandidate(int32 NumKeys, bool bHasDatas) { if (NumKeys > UE_HTTPDDC_BATCH_SIZE) { return false; } const uint32 Weight = ComputeWeight(NumKeys, bHasDatas); if (Weight > UE_HTTPDDC_BATCH_WEIGHT_HINT) { return false; } return true; } /** * Queues up a request to be batched. Blocks until the query is made. */ static FHttpRequest* QueueBatchRequest(FHttpRequestPool* InPool, const TCHAR* InNamespace, const TCHAR* InBucket, TConstArrayView InCacheKeys, TConstArrayView*> OutDatas, TBitArray<>& bOutVerified) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_BatchQuery); check(InCacheKeys.Num() == OutDatas.Num() || OutDatas.Num() == 0); const uint32 RequestNum = InCacheKeys.Num(); const uint32 RequestWeight = ComputeWeight(InCacheKeys.Num(), !OutDatas.IsEmpty()); for (int32 i = 0; i < Batches.Num(); i++) { uint32 Index = (FirstAvailableBatch.load(std::memory_order_relaxed) + i) % Batches.Num(); FBatch& Batch = Batches[Index]; //Assign different weights to head vs. get queries if (Batch.WeightHint.load(std::memory_order_acquire) + RequestWeight > UE_HTTPDDC_BATCH_WEIGHT_HINT) { continue; } // Attempt to reserve a spot in the batch const uint32 Reserve = Batch.Reserved.fetch_add(1, std::memory_order_acquire); if (Reserve >= UE_HTTPDDC_BATCH_SIZE) { // We didn't manage to snag a valid reserve index try next batch continue; } // Add our weight to the batch. Note we are treating it as a hint, so don't syncronize. const uint32 ActualWeight = Batch.WeightHint.fetch_add(RequestWeight, std::memory_order_release); TAnsiStringBuilder<64> BatchString; BatchString << "HttpDDC_Batch" << Index; TRACE_CPUPROFILER_EVENT_SCOPE_TEXT(*BatchString); if (Reserve == (UE_HTTPDDC_BATCH_SIZE - 1)) { FirstAvailableBatch++; } Batch.Entries[Reserve] = FQueuedBatchEntry{ InNamespace, InBucket, InCacheKeys, OutDatas, OutDatas.Num() ? FHttpRequest::ERequestVerb::Get : FHttpRequest::ERequestVerb::Head, &bOutVerified }; // Signal we are ready for batch to be submitted Batch.Ready.fetch_add(1u, std::memory_order_release); FHttpRequest* Request = nullptr; // The first to reserve a slot is the "driver" of the batch if (Reserve == 0) { Batch.Request = InPool->WaitForFreeRequest(); // Make sure no new requests are added const uint32 Reserved = FMath::Min((uint32)UE_HTTPDDC_BATCH_SIZE, Batch.Reserved.fetch_add(UE_HTTPDDC_BATCH_SIZE, std::memory_order_acquire)); // Give other threads time to copy their data to batch while (Batch.Ready.load(std::memory_order_acquire) < Reserved) { } // Increment request ref count to reflect all waiting threads InPool->MakeRequestShared(Batch.Request, Reserved); // Do the actual query and write response to respective target arrays PerformBatchQuery(Batch.Request, TArrayView(Batch.Entries, Batch.Ready)); // Signal to waiting threads the batch is complete Batch.Complete->Trigger(); // Store away the request and wait until other threads have too Request = Batch.Request; while (Batch.Ready.load(std::memory_order_acquire) > 1) { } //Reset batch for next use Batch.Complete->Reset(); Batch.WeightHint.store(0, std::memory_order_release); Batch.Ready.store(0, std::memory_order_release); Batch.Reserved.store(0, std::memory_order_release); } else { // Wait until "driver" has done query { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_WaitForMasterOfBatch); Batch.Complete->Wait(~0); } // Store away request and signal we are done Request = Batch.Request; Batch.Ready.fetch_sub(1u, std::memory_order_release); } return Request; } return nullptr; } /** * Creates request uri and headers and submits the request */ static void PerformBatchQuery(FHttpRequest* Request, TArrayView Entries) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_BatchGet); const TCHAR* Uri(TEXT("api/v1/c/ddc-rpc/batchget")); int64 ResponseCode = 0; uint32 Attempts = 0; //Prepare request object TArray> Operations; for (const FQueuedBatchEntry& Entry : Entries) { for (int32 KeyIdx = 0; KeyIdx < Entry.CacheKeys.Num(); KeyIdx++) { TSharedPtr Object = MakeShared(); Object->SetField(TEXT("bucket"), MakeShared(Entry.Bucket)); Object->SetField(TEXT("key"), MakeShared(Entry.CacheKeys[KeyIdx])); if (Entry.Verb == FHttpRequest::ERequestVerb::Head) { Object->SetField(TEXT("verb"), MakeShared(TEXT("HEAD"))); } Operations.Add(MakeShared(Object)); } } TSharedPtr RequestObject = MakeShared(); RequestObject->SetField(TEXT("namespace"), MakeShared(Entries[0].Namespace)); RequestObject->SetField(TEXT("operations"), MakeShared(Operations)); //Serialize to a buffer FBufferArchive RequestData; if (FJsonSerializer::Serialize(RequestObject.ToSharedRef(), TJsonWriterFactory>::Create(&RequestData))) { Request->PerformBlockingPost(Uri, FCompositeBuffer(FSharedBuffer::MakeView(RequestData.GetData(), RequestData.Num())), EHttpContentType::JSON); ResponseCode = Request->GetResponseCode(); if (ResponseCode == 200) { const TArray64& ResponseBuffer = Request->GetResponseBuffer(); const uint8* Response = ResponseBuffer.GetData(); const int32 ResponseSize = ResponseBuffer.Num(); // Parse the response and move the data to the target requests. if (ParseBatchedResponse(Response, ResponseSize, Entries)) { UE_LOG(LogDerivedDataCache, VeryVerbose, TEXT("%s: Batch query with %d operations completed."), *Request->GetName(), Entries.Num()); return; } } } // If we get here the request failed. UE_LOG(LogDerivedDataCache, Display, TEXT("%s: Batch query failed. Query: %s"), *Request->GetName(), ANSI_TO_TCHAR((ANSICHAR*)RequestData.GetData())); // Set all batch operations to failures for (FQueuedBatchEntry Entry : Entries) { Entry.bSuccess->SetRange(0, Entry.CacheKeys.Num(), false); } } // Above result value enum class OpResult : uint8 { Ok = 0, // Op finished succesfully Error = 1, // Error during op NotFound = 2, // Key was not found Exists = 3 // Used to indicate head op success }; // Searches for potentially multiple key requests that are satisfied the given cache key result // Search strategy is exhaustive forward search from the last found entry. If the results come in ordered the same as the requests, // and there are no duplicates, the search will be somewhat efficient (still has to do exhaustive searching looking for duplicates). // If the results are unordered or there are duplicates, search will become more inefficient. struct FRequestSearchHelper { FRequestSearchHelper(TArrayView InRequests, const FUTF8ToTCHAR& InCacheKey, int32 InEntryIdx, int32 InKeyIdx, OpResult InRequestResult) : Requests(InRequests) , CacheKey(InCacheKey) , StartEntryIdx(InEntryIdx) , StartKeyIdx(InKeyIdx) , RequestResult(InRequestResult) { } bool FindNext(int32& EntryIdx, int32& KeyIdx) { int32 CurrentEntryIdx = EntryIdx; int32 CurrentKeyIdx = KeyIdx; do { // Do not match a get request with a head response code (i.e. Exists) // or a head request with a get response code (i.e. Ok) // if the response code is an error or not found they can be matched to both head or get request it doesn't matter const FQueuedBatchEntry& CurrentRequest = Requests[CurrentEntryIdx]; bool bRequestTypeMatch = !((CurrentRequest.Verb == FHttpRequest::ERequestVerb::Get) && (RequestResult == OpResult::Exists)) && !((CurrentRequest.Verb == FHttpRequest::ERequestVerb::Head) && (RequestResult == OpResult::Ok)); if (bRequestTypeMatch && FCString::Stricmp(CurrentRequest.CacheKeys[CurrentKeyIdx], CacheKey.Get()) == 0) { EntryIdx = CurrentEntryIdx; KeyIdx = CurrentKeyIdx; return true; } } while (AdvanceIndices(CurrentEntryIdx, CurrentKeyIdx)); return false; } bool AdvanceIndices(int32& EntryIdx, int32& KeyIdx) { if (++KeyIdx >= Requests[EntryIdx].CacheKeys.Num()) { EntryIdx = (EntryIdx + 1) % Requests.Num(); KeyIdx = 0; } return !((EntryIdx == StartEntryIdx) && (KeyIdx == StartKeyIdx)); } TArrayView Requests; const FUTF8ToTCHAR& CacheKey; int32 StartEntryIdx; int32 StartKeyIdx; OpResult RequestResult; }; /** * Parses a batched response stream, moves the data to target requests and marks them with result. * @param Response Pointer to Response buffer * @param ResponseSize Size of response buffer * @param Requests Requests that will be filled with data. * @return True if response was successfully parsed, false otherwise. */ static bool ParseBatchedResponse(const uint8* ResponseStart, const int32 ResponseSize, TArrayView Requests) { // The expected data stream is structured accordingly // {"JPTR"} {PayloadCount:uint32} {{"JPEE"} {Name:cstr} {Result:uint8} {Hash:IoHash} {Size:uint64} {Payload...}} ... const auto& ResponseErrorMessage = TEXT("Malformed response from server."); const ANSICHAR* ProtocolMagic = "JPTR"; const ANSICHAR* PayloadMagic = "JPEE"; const uint32 MagicSize = 4; const uint8* Response = ResponseStart; const uint8* ResponseEnd = Response + ResponseSize; // Check that the stream starts with the protocol magic if (FMemory::Memcmp(ProtocolMagic, Response, MagicSize) != 0) { UE_LOG(LogDerivedDataCache, Display, ResponseErrorMessage); return false; } Response += MagicSize; // Number of payloads recieved uint32 PayloadCount = *(uint32*)Response; Response += sizeof(uint32); uint32 PayloadIdx = 0; // Current processed result int32 EntryIdx = 0; // Current Entry index int32 KeyIdx = 0; // Current Key index for current Entry while (Response < ResponseEnd && FMemory::Memcmp(PayloadMagic, Response, MagicSize) == 0) { PayloadIdx++; Response += MagicSize; const ANSICHAR* PayloadNameA = (const ANSICHAR*)Response; Response += FCStringAnsi::Strlen(PayloadNameA) + 1; //String and zero termination const ANSICHAR* CacheKeyA = FCStringAnsi::Strrchr(PayloadNameA, '.') + 1; // "namespace.bucket.cachekey" // Result of the operation is used to match to the appropriate request (i.e. get or head) OpResult PayloadResult = static_cast(*Response); Response += sizeof(uint8); const uint8* ResponseRewindMark = Response; // Find the payload among the requests. Payloads may be returned in any order and if the same cache key was part of two requests, // a single payload may satisfy multiple cache keys in multiple requests. FUTF8ToTCHAR CacheKey(CacheKeyA); FRequestSearchHelper RequestSearch(Requests, CacheKey, EntryIdx, KeyIdx, PayloadResult); bool bFoundAny = false; while (RequestSearch.FindNext(EntryIdx, KeyIdx)) { Response = ResponseRewindMark; bFoundAny = true; FQueuedBatchEntry& RequestOp = Requests[EntryIdx]; TBitArray<>& bSuccess = *RequestOp.bSuccess; switch (PayloadResult) { case OpResult::Ok: { // Payload hash of the following payload data FIoHash PayloadHash = *(FIoHash*)Response; Response += sizeof(FIoHash); // Size of the following payload data const uint64 PayloadSize = *(uint64*)Response; Response += sizeof(uint64); if (PayloadSize > 0) { if (Response + PayloadSize > ResponseEnd) { UE_LOG(LogDerivedDataCache, Display, ResponseErrorMessage); return false; } if (bSuccess[KeyIdx]) { Response += PayloadSize; } else { TArray64* OutData = RequestOp.OutDatas[KeyIdx]; OutData->Append(Response, PayloadSize); Response += PayloadSize; // Verify the received and parsed payload if (VerifyPayload(PayloadHash, RequestOp.Namespace, RequestOp.Bucket, RequestOp.CacheKeys[KeyIdx], *OutData)) { TRACE_COUNTER_ADD(HttpDDC_GetHit, int64(1)); TRACE_COUNTER_ADD(HttpDDC_BytesReceived, int64(PayloadSize)); bSuccess[KeyIdx] = true; } else { OutData->Empty(); bSuccess[KeyIdx] = false; } } } else { bSuccess[KeyIdx] = false; } } break; case OpResult::Exists: { TRACE_COUNTER_ADD(HttpDDC_ExistHit, int64(1)); bSuccess[KeyIdx] = true; } break; default: case OpResult::Error: UE_LOG(LogDerivedDataCache, Display, TEXT("Server error while getting %s"), CacheKey.Get()); // intentional falltrough case OpResult::NotFound: bSuccess[KeyIdx] = false; break; } if (!RequestSearch.AdvanceIndices(EntryIdx, KeyIdx)) { break; } } if (!bFoundAny) { UE_LOG(LogDerivedDataCache, Error, ResponseErrorMessage); return false; } } // Have we parsed all the payloads from the message? if (PayloadIdx != PayloadCount) { UE_LOG(LogDerivedDataCache, Display, TEXT("%s: Found %d payloads but %d was reported."), ResponseErrorMessage, PayloadIdx, PayloadCount); } return true; } }; TStaticArray FDataRequestHelper::Batches; std::atomic FDataRequestHelper::FirstAvailableBatch; //---------------------------------------------------------------------------------------------------------- // FDataUploadHelper //---------------------------------------------------------------------------------------------------------- struct FDataUploadHelper { FDataUploadHelper(FHttpRequestPool* InPool, const TCHAR* InNamespace, const TCHAR* InBucket, const TCHAR* InCacheKey, const TArrayView& InData, FDerivedDataCacheUsageStats& InUsageStats) : ResponseCode(0) , bSuccess(false) , bQueued(false) { FHttpRequest* Request = InPool->GetFreeRequest(); if (Request) { ResponseCode = PerformPut(Request, InNamespace, InBucket, InCacheKey, InData, InUsageStats); bSuccess = FHttpRequest::IsSuccessResponse(Request->GetResponseCode()); ProcessQueuedPutsAndReleaseRequest(InPool, Request, InUsageStats); } else { FQueuedEntry* Entry = new FQueuedEntry(InNamespace, InBucket, InCacheKey, InData); QueuedPuts.Push(Entry); bSuccess = true; bQueued = true; // A request may have been released while the entry was being queued. Request = InPool->GetFreeRequest(); if (Request) { ProcessQueuedPutsAndReleaseRequest(InPool, Request, InUsageStats); } } } bool IsSuccess() const { return bSuccess; } int64 GetResponseCode() const { return ResponseCode; } bool IsQueued() const { return bQueued; } private: struct FQueuedEntry { FString Namespace; FString Bucket; FString CacheKey; TArray64 Data; FQueuedEntry(const TCHAR* InNamespace, const TCHAR* InBucket, const TCHAR* InCacheKey, const TArrayView InData) : Namespace(InNamespace) , Bucket(InBucket) , CacheKey(InCacheKey) , Data(InData) // Copies the data! { } }; static TLockFreePointerListUnordered QueuedPuts; int64 ResponseCode; bool bSuccess; bool bQueued; static void ProcessQueuedPutsAndReleaseRequest(FHttpRequestPool* Pool, FHttpRequest* Request, FDerivedDataCacheUsageStats& UsageStats) { while (Request) { // Make sure that whether we early exit or execute past the end of this scope that // the request is released back to the pool. { ON_SCOPE_EXIT { Pool->ReleaseRequestToPool(Request); }; if (ShouldAbortForShutdown()) { return; } while (FQueuedEntry* Entry = QueuedPuts.Pop()) { Request->Reset(); PerformPut(Request, *Entry->Namespace, *Entry->Bucket, *Entry->CacheKey, Entry->Data, UsageStats); delete Entry; if (ShouldAbortForShutdown()) { return; } } } // An entry may have been queued while the request was being released. if (QueuedPuts.IsEmpty()) { break; } // Process the queue again if a request is free, otherwise the thread that got the request will process it. Request = Pool->GetFreeRequest(); } } static int64 PerformPut(FHttpRequest* Request, const TCHAR* Namespace, const TCHAR* Bucket, const TCHAR* CacheKey, const TArrayView Data, FDerivedDataCacheUsageStats& UsageStats) { COOK_STAT(auto Timer = UsageStats.TimePut()); HashPayload(Request, Data); TStringBuilder<256> Uri; Uri.Appendf(TEXT("api/v1/c/ddc/%s/%s/%s"), Namespace, Bucket, CacheKey); Request->PerformBlockingPut(*Uri, FCompositeBuffer(FSharedBuffer::MakeView(Data.GetData(), Data.Num())), EHttpContentType::Binary); const int64 ResponseCode = Request->GetResponseCode(); if (FHttpRequest::IsSuccessResponse(ResponseCode)) { TRACE_COUNTER_ADD(HttpDDC_BytesSent, int64(Request->GetBytesSent())); COOK_STAT(Timer.AddHit(Request->GetBytesSent())); } return Request->GetResponseCode(); } }; TLockFreePointerListUnordered FDataUploadHelper::QueuedPuts; #endif // WITH_DATAREQUEST_HELPER //---------------------------------------------------------------------------------------------------------- // Content parsing and checking //---------------------------------------------------------------------------------------------------------- /** * Verifies the integrity of the received data using supplied checksum. * @param Hash received hash value. * @param Namespace The namespace string used when originally fetching the request. * @param Bucket The bucket string used when originally fetching the request. * @param CacheKey The cache key string used when originally fetching the request. * @param Payload Payload received. * @return True if the data is correct, false if checksums doesn't match. */ bool VerifyPayload(const FSHAHash& Hash, const TCHAR* Namespace, const TCHAR* Bucket, const TCHAR* CacheKey, const TArray64& Payload) { FSHAHash PayloadHash; FSHA1::HashBuffer(Payload.GetData(), Payload.Num(), PayloadHash.Hash); if (Hash != PayloadHash) { UE_LOG(LogDerivedDataCache, Display, TEXT("Checksum from server did not match received data (%s vs %s). Discarding cached result. Namespace: %s, Bucket: %s, Key: %s."), *WriteToString<48>(Hash), *WriteToString<48>(PayloadHash), Namespace, Bucket, CacheKey ); return false; } return true; } /** * Verifies the integrity of the received data using supplied checksum. * @param Hash received hash value. * @param Namespace The namespace string used when originally fetching the request. * @param Bucket The bucket string used when originally fetching the request. * @param CacheKey The cache key string used when originally fetching the request. * @param Payload Payload received. * @return True if the data is correct, false if checksums doesn't match. */ bool VerifyPayload(const FIoHash& Hash, const TCHAR* Namespace, const TCHAR* Bucket, const TCHAR* CacheKey, const TArray64& Payload) { FIoHash PayloadHash = FIoHash::HashBuffer(Payload.GetData(), Payload.Num()); if (Hash != PayloadHash) { UE_LOG(LogDerivedDataCache, Display, TEXT("Checksum from server did not match received data (%s vs %s). Discarding cached result. Namespace: %s, Bucket: %s, Key: %s."), *WriteToString<48>(Hash), *WriteToString<48>(PayloadHash), Namespace, Bucket, CacheKey ); return false; } return true; } /** * Verifies the integrity of the received data using supplied checksum. * @param Request Request that the data was be received with. * @param Namespace The namespace string used when originally fetching the request. * @param Bucket The bucket string used when originally fetching the request. * @param CacheKey The cache key string used when originally fetching the request. * @param Payload Payload received. * @return True if the data is correct, false if checksums doesn't match. */ bool VerifyRequest(const FHttpRequest* Request, const TCHAR* Namespace, const TCHAR* Bucket, const TCHAR* CacheKey, const TArray64& Payload) { FString ReceivedHashStr; if (Request->GetHeader("X-Jupiter-Sha1", ReceivedHashStr)) { FSHAHash ReceivedHash; ReceivedHash.FromString(ReceivedHashStr); return VerifyPayload(ReceivedHash, Namespace, Bucket, CacheKey, Payload); } if (Request->GetHeader("X-Jupiter-IoHash", ReceivedHashStr)) { FIoHash ReceivedHash(ReceivedHashStr); return VerifyPayload(ReceivedHash, Namespace, Bucket, CacheKey, Payload); } UE_LOG(LogDerivedDataCache, Warning, TEXT("%s: HTTP server did not send a content hash. Wrong server version?"), *Request->GetName()); return true; } /** * Adds a checksum (as request header) for a given payload. Jupiter will use this to verify the integrity * of the received data. * @param Request Request that the data will be sent with. * @param Payload Payload that will be sent. * @return True on success, false on failure. */ bool HashPayload(FHttpRequest* Request, const TArrayView Payload) { FIoHash PayloadHash = FIoHash::HashBuffer(Payload.GetData(), Payload.Num()); Request->AddHeader(TEXTVIEW("X-Jupiter-IoHash"), WriteToString<48>(PayloadHash)); return true; } bool ShouldAbortForShutdown() { return !GIsBuildMachine && FDerivedDataBackend::Get().IsShuttingDown(); } TConstArrayView MakeConstArrayView(FSharedBuffer Buffer) { return TConstArrayView(reinterpret_cast(Buffer.GetData()), Buffer.GetSize()); } static bool IsValueDataReady(FValue& Value, const ECachePolicy Policy) { if (!EnumHasAnyFlags(Policy, ECachePolicy::Query)) { Value = Value.RemoveData(); return true; } if (Value.HasData()) { if (EnumHasAnyFlags(Policy, ECachePolicy::SkipData)) { Value = Value.RemoveData(); } return true; } return false; }; //---------------------------------------------------------------------------------------------------------- // FHttpCacheStore //---------------------------------------------------------------------------------------------------------- /** * Backend for a HTTP based caching service (Jupiter). **/ class FHttpCacheStore final : public FDerivedDataBackendInterface { public: /** * Creates the backend, checks health status and attempts to acquire an access token. * * @param ServiceUrl Base url to the service including schema. * @param Namespace Namespace to use. * @param StructuredNamespace Namespace to use for structured cache operations. * @param OAuthProvider Url to OAuth provider, for example "https://myprovider.com/oauth2/v1/token". * @param OAuthClientId OAuth client identifier. * @param OAuthData OAuth form data to send to login service. Can either be the raw form data or a Windows network file address (starting with "\\"). * @param OAuthScope OAuth scope identifier */ FHttpCacheStore( const TCHAR* ServiceUrl, bool bResolveHostCanonicalName, const TCHAR* Namespace, const TCHAR* StructuredNamespace, const TCHAR* OAuthProvider, const TCHAR* OAuthClientId, const TCHAR* OAuthData, const TCHAR* OAuthScope, EBackendLegacyMode LegacyMode, bool bReadOnly); ~FHttpCacheStore(); /** * Checks is backend is usable (reachable and accessible). * @return true if usable */ bool IsUsable() const { return bIsUsable; } /** return true if this cache is writable **/ virtual bool IsWritable() const override { return !bReadOnly && bIsUsable; } virtual bool CachedDataProbablyExists(const TCHAR* CacheKey) override; virtual TBitArray<> CachedDataProbablyExistsBatch(TConstArrayView CacheKeys) override; virtual bool GetCachedData(const TCHAR* CacheKey, TArray& OutData) override; virtual EPutStatus PutCachedData(const TCHAR* CacheKey, TArrayView InData, bool bPutEvenIfExists) override; virtual void RemoveCachedData(const TCHAR* CacheKey, bool bTransient) override; virtual TSharedRef GatherUsageStats() const override; virtual FString GetName() const override; virtual TBitArray<> TryToPrefetch(TConstArrayView CacheKeys) override; virtual bool WouldCache(const TCHAR* CacheKey, TArrayView InData) override; virtual ESpeedClass GetSpeedClass() const override; virtual bool ApplyDebugOptions(FBackendDebugOptions& InOptions) override; void SetSpeedClass(ESpeedClass InSpeedClass) { SpeedClass = InSpeedClass; } EBackendLegacyMode GetLegacyMode() const final { return LegacyMode; } virtual void Put( TConstArrayView Requests, IRequestOwner& Owner, FOnCachePutComplete&& OnComplete) override; virtual void Get( TConstArrayView Requests, IRequestOwner& Owner, FOnCacheGetComplete&& OnComplete) override; virtual void PutValue( TConstArrayView Requests, IRequestOwner& Owner, FOnCachePutValueComplete&& OnComplete) override; virtual void GetValue( TConstArrayView Requests, IRequestOwner& Owner, FOnCacheGetValueComplete&& OnComplete) override; virtual void GetChunks( TConstArrayView Requests, IRequestOwner& Owner, FOnCacheGetChunkComplete&& OnComplete) override; static FHttpCacheStore* GetAny() { return AnyInstance; } const FString& GetDomain() const { return Domain; } const FString& GetNamespace() const { return Namespace; } const FString& GetStructuredNamespace() const { return StructuredNamespace; } const FString& GetOAuthProvider() const { return OAuthProvider; } const FString& GetOAuthClientId() const { return OAuthClientId; } const FString& GetOAuthSecret() const { return OAuthSecret; } const FString& GetOAuthScope() const { return OAuthScope; } private: FString Domain; FString EffectiveDomain; FString Namespace; FString StructuredNamespace; FString DefaultBucket; FString OAuthProvider; FString OAuthClientId; FString OAuthSecret; FString OAuthScope; FCriticalSection AccessCs; FDerivedDataCacheUsageStats UsageStats; FBackendDebugOptions DebugOptions; TUniquePtr SharedData; TUniquePtr GetRequestPools[2]; TUniquePtr PutRequestPools[2]; TUniquePtr NonBlockingRequestPools; TUniquePtr Access; bool bIsUsable; bool bReadOnly; uint32 FailedLoginAttempts; ESpeedClass SpeedClass; EBackendLegacyMode LegacyMode; static inline FHttpCacheStore* AnyInstance = nullptr; bool IsServiceReady(); bool AcquireAccessToken(); bool ShouldRetryOnError(FHttpRequest::EResult Result, int64 ResponseCode); bool ShouldRetryOnError(int64 ResponseCode) { return ShouldRetryOnError(FHttpRequest::EResult::Success, ResponseCode); } enum class OperationCategory { Get, Put }; template FHttpRequest* WaitForHttpRequestForOwner(IRequestOwner& Owner, bool bUnboundedOverflow, FHttpRequestPool*& OutPool) { if (!FHttpRequest::AllowAsync()) { if (Category == OperationCategory::Get) { OutPool = GetRequestPools[IsInGameThread()].Get(); } else { OutPool = PutRequestPools[IsInGameThread()].Get(); } return OutPool->WaitForFreeRequest(); } else { OutPool = NonBlockingRequestPools.Get(); return OutPool->WaitForFreeRequest(bUnboundedOverflow); } } struct FGetCacheRecordOnlyResponse { FSharedString Name; FCacheKey Key; uint64 UserData = 0; uint64 BytesReceived = 0; FOptionalCacheRecord Record; EStatus Status = EStatus::Error; }; using FOnGetCacheRecordOnlyComplete = TUniqueFunction; void GetCacheRecordOnlyAsync( IRequestOwner& Owner, const FSharedString& Name, const FCacheKey& Key, const FCacheRecordPolicy& Policy, uint64 UserData, FOnGetCacheRecordOnlyComplete&& OnComplete); void GetCacheRecordAsync( IRequestOwner& Owner, const FSharedString& Name, const FCacheKey& Key, const FCacheRecordPolicy& Policy, uint64 UserData, TUniqueFunction&& OnComplete); void PutCacheRecordAsync( IRequestOwner& Owner, const FSharedString& Name, const FCacheRecord& Record, const FCacheRecordPolicy& Policy, uint64 UserData, TUniqueFunction&& OnComplete); void PutCacheValueAsync( IRequestOwner& Owner, const FSharedString& Name, const FCacheKey& Key, const FValue& Value, ECachePolicy Policy, uint64 UserData, TUniqueFunction&& OnComplete); void GetCacheValueAsync( IRequestOwner& Owner, FSharedString Name, const FCacheKey& Key, ECachePolicy Policy, uint64 UserData, FOnCacheGetValueComplete&& OnComplete); void RefCachedDataProbablyExistsBatchAsync( IRequestOwner& Owner, TConstArrayView ValueRefs, FOnCacheGetValueComplete&& OnComplete); class FPutPackageOp; class FGetRecordOp; }; //---------------------------------------------------------------------------------------------------------- // FHttpCacheStore::FPutPackageOp //---------------------------------------------------------------------------------------------------------- class FHttpCacheStore::FPutPackageOp final : public FThreadSafeRefCountedObject { public: struct FCachePutPackageResponse { FSharedString Name; FCacheKey Key; uint64 UserData = 0; uint64 BytesSent = 0; EStatus Status = EStatus::Error; }; using FOnCachePutPackageComplete = TUniqueFunction; /** Performs a multi-request operation for uploading a package of content. */ static void PutPackage( FHttpCacheStore& CacheStore, IRequestOwner& Owner, const FSharedString& Name, FCacheKey Key, FCbPackage&& Package, FCacheRecordPolicy Policy, uint64 UserData, FOnCachePutPackageComplete&& OnComplete); private: FHttpCacheStore& CacheStore; IRequestOwner& Owner; const FSharedString Name; const FCacheKey Key; const uint64 UserData; std::atomic BytesSent; const FCbObject PackageObject; const FIoHash PackageObjectHash; const uint32 TotalBlobUploads; std::atomic SuccessfulBlobUploads; std::atomic PendingBlobUploads; FOnCachePutPackageComplete OnComplete; struct FCachePutRefResponse { FSharedString Name; FCacheKey Key; uint64 UserData = 0; uint64 BytesSent = 0; TConstArrayView NeededBlobHashes; EStatus Status = EStatus::Error; }; using FOnCachePutRefComplete = TUniqueFunction; FPutPackageOp( FHttpCacheStore& InCacheStore, IRequestOwner& InOwner, const FSharedString& InName, const FCacheKey& InKey, uint64 InUserData, uint64 InBytesSent, const FCbObject& InPackageObject, const FIoHash& InPackageObjectHash, uint32 InTotalBlobUploads, FOnCachePutPackageComplete&& InOnComplete); static void PutRefAsync( FHttpCacheStore& CacheStore, IRequestOwner& Owner, const FSharedString& Name, FCacheKey Key, FCbObject Object, FIoHash ObjectHash, uint64 UserData, bool bFinalize, FOnCachePutRefComplete&& OnComplete); static void OnPackagePutRefComplete( FHttpCacheStore& CacheStore, IRequestOwner& Owner, const FSharedString& Name, const FCacheKey& Key, FCbPackage&& Package, FCacheRecordPolicy Policy, uint64 UserData, FOnCachePutPackageComplete&& OnComplete, FCachePutRefResponse&& Response); FHttpRequest::ECompletionBehavior OnCompressedBlobUploadComplete( FHttpRequest::EResult HttpResult, FHttpRequest* Request); void OnPutRefFinalizationComplete( FCachePutRefResponse&& Response); FCachePutPackageResponse MakeResponse(uint64 InBytesSent, EStatus Status) { return FCachePutPackageResponse{ Name, Key, UserData, InBytesSent, Status }; }; }; //---------------------------------------------------------------------------------------------------------- // FHttpCacheStore::FGetRecordOp //---------------------------------------------------------------------------------------------------------- class FHttpCacheStore::FGetRecordOp final : public FThreadSafeRefCountedObject { public: /** Performs a multi-request operation for downloading a record. */ static void GetRecord( FHttpCacheStore& CacheStore, IRequestOwner& Owner, const FSharedString& Name, const FCacheKey& Key, const FCacheRecordPolicy& Policy, uint64 UserData, TUniqueFunction&& OnComplete); struct FGetCachedDataBatchResponse { FSharedString Name; FCacheKey Key; int32 ValueIndex; uint64 BytesReceived = 0; FCompressedBuffer DataBuffer; EStatus Status = EStatus::Error; }; using FOnGetCachedDataBatchComplete = TUniqueFunction; /** Utility method for fetching a batch of value data. */ template static void GetDataBatch( FHttpCacheStore& CacheStore, IRequestOwner& Owner, FSharedString Name, const FCacheKey& Key, TConstArrayView Values, ValueIdGetterType ValueIdGetter, FOnGetCachedDataBatchComplete&& OnComplete); private: FHttpCacheStore& CacheStore; IRequestOwner& Owner; const FSharedString Name; const FCacheKey Key; const uint64 UserData; std::atomic BytesReceived; TArray FetchedBuffers; const TArray RequiredGets; const TArray RequiredHeads; FCacheRecordBuilder RecordBuilder; const uint32 TotalOperations; std::atomic SuccessfulOperations; std::atomic PendingOperations; TUniqueFunction OnComplete; FGetRecordOp( FHttpCacheStore& InCacheStore, IRequestOwner& InOwner, const FSharedString& InName, const FCacheKey& InKey, uint64 InUserData, uint64 InBytesReceived, TArray&& InRequiredGets, TArray&& InRequiredHeads, FCacheRecordBuilder&& InRecordBuilder, TUniqueFunction&& InOnComplete); static void OnOnlyRecordComplete( FHttpCacheStore& CacheStore, IRequestOwner& Owner, const FCacheRecordPolicy& Policy, TUniqueFunction&& OnComplete, FGetCacheRecordOnlyResponse&& Response); struct FCachedDataProbablyExistsBatchResponse { FSharedString Name; FCacheKey Key; int32 ValueIndex; EStatus Status = EStatus::Error; }; using FOnCachedDataProbablyExistsBatchComplete = TUniqueFunction; void DataProbablyExistsBatch( TConstArrayView Values, FOnCachedDataProbablyExistsBatchComplete&& OnComplete); void FinishDataStep(bool bSuccess, uint64 InBytesReceived); }; void FHttpCacheStore::FPutPackageOp::PutPackage( FHttpCacheStore& CacheStore, IRequestOwner& Owner, const FSharedString& Name, FCacheKey Key, FCbPackage&& Package, FCacheRecordPolicy Policy, uint64 UserData, FOnCachePutPackageComplete&& OnComplete) { // TODO: Jupiter currently always overwrites. It doesn't have a "write if not present" feature (for records or attachments), // but would require one to implement all policy correctly. // Initial record upload PutRefAsync(CacheStore, Owner, Name, Key, Package.GetObject(), Package.GetObjectHash(), UserData, false, [&CacheStore, &Owner, Name = FSharedString(Name), Key, Package = MoveTemp(Package), Policy, UserData, OnComplete = MoveTemp(OnComplete)](FCachePutRefResponse&& Response) mutable { return OnPackagePutRefComplete(CacheStore, Owner, Name, Key, MoveTemp(Package), Policy, UserData, MoveTemp(OnComplete), MoveTemp(Response)); }); } FHttpCacheStore::FPutPackageOp::FPutPackageOp( FHttpCacheStore& InCacheStore, IRequestOwner& InOwner, const FSharedString& InName, const FCacheKey& InKey, uint64 InUserData, uint64 InBytesSent, const FCbObject& InPackageObject, const FIoHash& InPackageObjectHash, uint32 InTotalBlobUploads, FOnCachePutPackageComplete&& InOnComplete) : CacheStore(InCacheStore) , Owner(InOwner) , Name(InName) , Key(InKey) , UserData(InUserData) , BytesSent(InBytesSent) , PackageObject(InPackageObject) , PackageObjectHash(InPackageObjectHash) , TotalBlobUploads(InTotalBlobUploads) , SuccessfulBlobUploads(0) , PendingBlobUploads(InTotalBlobUploads) , OnComplete(MoveTemp(InOnComplete)) { } void FHttpCacheStore::FPutPackageOp::PutRefAsync( FHttpCacheStore& CacheStore, IRequestOwner& Owner, const FSharedString& Name, FCacheKey Key, FCbObject Object, FIoHash ObjectHash, uint64 UserData, bool bFinalize, FOnCachePutRefComplete&& OnComplete) { FString Bucket(Key.Bucket.ToString()); Bucket.ToLowerInline(); TStringBuilder<256> RefsUri; RefsUri << "api/v1/refs/" << CacheStore.StructuredNamespace << "/" << Bucket << "/" << Key.Hash; if (bFinalize) { RefsUri << "/finalize/" << ObjectHash; } FHttpRequestPool* Pool = nullptr; FHttpRequest* Request = CacheStore.WaitForHttpRequestForOwner(Owner, bFinalize /* bUnboundedOverflow */, Pool); auto OnHttpRequestComplete = [&CacheStore, &Owner, Name = FSharedString(Name), Key, Object, UserData, bFinalize, OnComplete = MoveTemp(OnComplete)] (FHttpRequest::EResult HttpResult, FHttpRequest* Request) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_PutRefAsync_OnHttpRequestComplete); if (Owner.IsCanceled()) { OnComplete({ Name, Key, UserData, Request->GetBytesSent(), {}, EStatus::Canceled }); return FHttpRequest::ECompletionBehavior::Done; } int64 ResponseCode = Request->GetResponseCode(); if (FHttpRequest::IsSuccessResponse(ResponseCode)) { TArray NeededBlobHashes; // Useful when debugging issues related to compressed/uncompressed blobs being returned from Jupiter const bool bPutRefBlobsAlways = false; if (bPutRefBlobsAlways && !bFinalize) { Object.IterateAttachments([&NeededBlobHashes](FCbFieldView AttachmentFieldView) { FIoHash AttachmentHash = AttachmentFieldView.AsHash(); if (!AttachmentHash.IsZero()) { NeededBlobHashes.Add(AttachmentHash); } }); } else if (TSharedPtr ResponseObject = Request->GetResponseAsJsonObject()) { TArray NeedsArrayStrings; ResponseObject->TryGetStringArrayField(TEXT("needs"), NeedsArrayStrings); NeededBlobHashes.Reserve(NeedsArrayStrings.Num()); for (const FString& NeededString : NeedsArrayStrings) { FIoHash BlobHash; LexFromString(BlobHash, *NeededString); if (!BlobHash.IsZero()) { NeededBlobHashes.Add(BlobHash); } } } OnComplete({ Name, Key, UserData, Request->GetBytesSent(), NeededBlobHashes, EStatus::Ok }); return FHttpRequest::ECompletionBehavior::Done; } if (!ShouldAbortForShutdown() && CacheStore.ShouldRetryOnError(HttpResult, ResponseCode) && ((Request->GetAttempts()+1) < UE_HTTPDDC_MAX_ATTEMPTS)) { return FHttpRequest::ECompletionBehavior::Retry; } OnComplete({ Name, Key, UserData, Request->GetBytesSent(), {}, EStatus::Error }); return FHttpRequest::ECompletionBehavior::Done; }; if (bFinalize) { Request->EnqueueAsyncPost(Owner, Pool, *RefsUri, FCompositeBuffer(), MoveTemp(OnHttpRequestComplete), EHttpContentType::FormUrlEncoded); } else { Request->AddHeader(TEXTVIEW("X-Jupiter-IoHash"), WriteToString<48>(ObjectHash)); Request->EnqueueAsyncPut(Owner, Pool, *RefsUri, Object.GetBuffer(), MoveTemp(OnHttpRequestComplete), EHttpContentType::CbObject); } } void FHttpCacheStore::FPutPackageOp::OnPackagePutRefComplete( FHttpCacheStore& CacheStore, IRequestOwner& Owner, const FSharedString& Name, const FCacheKey& Key, FCbPackage&& Package, FCacheRecordPolicy Policy, uint64 UserData, FOnCachePutPackageComplete&& OnComplete, FCachePutRefResponse&& Response) { if (Response.Status != EStatus::Ok) { if (Response.Status == EStatus::Error) { UE_LOG(LogDerivedDataCache, Log, TEXT("%s: Failed to put reference object for put of %s from '%s'"), *CacheStore.GetName(), *WriteToString<96>(Response.Key), *Response.Name); } return OnComplete(FCachePutPackageResponse{ Name, Key, UserData, Response.BytesSent, Response.Status }); } struct FCompressedBlobUpload { FIoHash Hash; FSharedBuffer BlobBuffer; FCompressedBlobUpload(const FIoHash& InHash, FSharedBuffer&& InBlobBuffer) : Hash(InHash), BlobBuffer(InBlobBuffer) { } }; TArray CompressedBlobUploads; // TODO: blob uploading and finalization should be replaced with a single batch compressed blob upload endpoint in the future. TStringBuilder<128> ExpectedHashes; bool bExpectedHashesSerialized = false; // Needed blob upload (if any missing) for (const FIoHash& NeededBlobHash : Response.NeededBlobHashes) { if (const FCbAttachment* Attachment = Package.FindAttachment(NeededBlobHash)) { FSharedBuffer TempBuffer; if (Attachment->IsCompressedBinary()) { TempBuffer = Attachment->AsCompressedBinary().GetCompressed().ToShared(); } else if (Attachment->IsBinary()) { TempBuffer = FValue::Compress(Attachment->AsCompositeBinary()).GetData().GetCompressed().ToShared(); } else { TempBuffer = FValue::Compress(Attachment->AsObject().GetBuffer()).GetData().GetCompressed().ToShared(); } CompressedBlobUploads.Emplace(NeededBlobHash, MoveTemp(TempBuffer)); } else { if (!bExpectedHashesSerialized) { bool bFirstHash = true; for (const FCbAttachment& PackageAttachment : Package.GetAttachments()) { if (!bFirstHash) { ExpectedHashes << TEXT(", "); } ExpectedHashes << PackageAttachment.GetHash(); bFirstHash = false; } bExpectedHashesSerialized = true; } UE_LOG(LogDerivedDataCache, Log, TEXT("%s: Server reported needed hash '%s' that is outside the set of expected hashes (%s) for put of %s from '%s'"), *CacheStore.GetName(), *WriteToString<96>(NeededBlobHash), ExpectedHashes.ToString(), *WriteToString<96>(Response.Key), *Response.Name); } } if (CompressedBlobUploads.IsEmpty()) { // No blobs need to be uploaded. No finalization necessary. return OnComplete(FCachePutPackageResponse{ Name, Key, UserData, Response.BytesSent, EStatus::Ok }); } // Having this be a ref ensures we don't have the op reach 0 ref count as we queue up multiple operations which MAY execute synchronously TRefCountPtr PutPackageOp = new FPutPackageOp( CacheStore, Owner, Response.Name, Response.Key, Response.UserData, Response.BytesSent, Package.GetObject(), Package.GetObjectHash(), (uint32)CompressedBlobUploads.Num(), MoveTemp(OnComplete) ); FRequestBarrier Barrier(Owner); for (const FCompressedBlobUpload& CompressedBlobUpload : CompressedBlobUploads) { TStringBuilder<256> CompressedBlobsUri; CompressedBlobsUri << "api/v1/compressed-blobs/" << CacheStore.StructuredNamespace << "/" << CompressedBlobUpload.Hash; FHttpRequestPool* Pool = nullptr; FHttpRequest* Request = CacheStore.WaitForHttpRequestForOwner(Owner, true /* bUnboundedOverflow */, Pool); Request->EnqueueAsyncPut(Owner, Pool, *CompressedBlobsUri, FCompositeBuffer(CompressedBlobUpload.BlobBuffer), [PutPackageOp](FHttpRequest::EResult HttpResult, FHttpRequest* Request) { return PutPackageOp->OnCompressedBlobUploadComplete(HttpResult, Request); }, EHttpContentType::CompressedBinary); } } FHttpRequest::ECompletionBehavior FHttpCacheStore::FPutPackageOp::OnCompressedBlobUploadComplete( FHttpRequest::EResult HttpResult, FHttpRequest* Request) { int64 ResponseCode = Request->GetResponseCode(); bool bIsSuccessResponse = FHttpRequest::IsSuccessResponse(ResponseCode); if (!bIsSuccessResponse && !ShouldAbortForShutdown() && !Owner.IsCanceled() && CacheStore.ShouldRetryOnError(HttpResult, ResponseCode) && ((Request->GetAttempts()+1) < UE_HTTPDDC_MAX_ATTEMPTS)) { return FHttpRequest::ECompletionBehavior::Retry; } BytesSent.fetch_add(Request->GetBytesSent(), std::memory_order_relaxed); if (bIsSuccessResponse) { SuccessfulBlobUploads.fetch_add(1, std::memory_order_relaxed); } if (PendingBlobUploads.fetch_sub(1, std::memory_order_relaxed) == 1) { if (Owner.IsCanceled()) { OnComplete(MakeResponse(BytesSent.load(std::memory_order_relaxed), EStatus::Canceled)); return FHttpRequest::ECompletionBehavior::Done; } uint32 LocalSuccessfulBlobUploads = SuccessfulBlobUploads.load(std::memory_order_relaxed); if (LocalSuccessfulBlobUploads == TotalBlobUploads) { // Perform finalization PutRefAsync(CacheStore, Owner, Name, Key, PackageObject, PackageObjectHash, UserData, true, [PutPackageOp = TRefCountPtr(this)](FCachePutRefResponse&& Response) { return PutPackageOp->OnPutRefFinalizationComplete(MoveTemp(Response)); }); } else { uint32 FailedBlobUploads = (uint32)(TotalBlobUploads - LocalSuccessfulBlobUploads); UE_LOG(LogDerivedDataCache, Log, TEXT("%s: Failed to put %d/%d blobs for put of %s from '%s'"), *CacheStore.GetName(), FailedBlobUploads, TotalBlobUploads, *WriteToString<96>(Key), *Name); OnComplete(MakeResponse(BytesSent.load(std::memory_order_relaxed), EStatus::Error)); } } return FHttpRequest::ECompletionBehavior::Done; } void FHttpCacheStore::FPutPackageOp::OnPutRefFinalizationComplete( FCachePutRefResponse&& Response) { BytesSent.fetch_add(Response.BytesSent, std::memory_order_relaxed); if (Response.Status == EStatus::Error) { UE_LOG(LogDerivedDataCache, Log, TEXT("%s: Failed to finalize reference object for put of %s from '%s'"), *CacheStore.GetName(), *WriteToString<96>(Key), *Name); } return OnComplete(MakeResponse(BytesSent.load(std::memory_order_relaxed), Response.Status)); } void FHttpCacheStore::FGetRecordOp::GetRecord( FHttpCacheStore& CacheStore, IRequestOwner& Owner, const FSharedString& Name, const FCacheKey& Key, const FCacheRecordPolicy& Policy, uint64 UserData, TUniqueFunction&& OnComplete) { CacheStore.GetCacheRecordOnlyAsync(Owner, Name, Key, Policy, UserData, [&CacheStore, &Owner, Policy = FCacheRecordPolicy(Policy), OnComplete = MoveTemp(OnComplete)](FGetCacheRecordOnlyResponse&& Response) mutable { OnOnlyRecordComplete(CacheStore, Owner, Policy, MoveTemp(OnComplete), MoveTemp(Response)); }); } template void FHttpCacheStore::FGetRecordOp::GetDataBatch( FHttpCacheStore& CacheStore, IRequestOwner& Owner, FSharedString Name, const FCacheKey& Key, TConstArrayView Values, ValueIdGetterType ValueIdGetter, FOnGetCachedDataBatchComplete&& OnComplete) { if (Values.IsEmpty()) { return; } FRequestBarrier Barrier(Owner); TRefCountedUniqueFunction* CompletionFunction = new TRefCountedUniqueFunction(MoveTemp(OnComplete)); TRefCountPtr> BatchOnCompleteRef(CompletionFunction); for (int32 ValueIndex = 0; ValueIndex < Values.Num(); ++ValueIndex) { const ValueType& Value = Values[ValueIndex]; const FIoHash& RawHash = Value.GetRawHash(); FHttpRequestPool* Pool = nullptr; FHttpRequest* Request = CacheStore.WaitForHttpRequestForOwner(Owner, true /* bUnboundedOverflow */, Pool); auto OnHttpRequestComplete = [&CacheStore, &Owner, Name = FSharedString(Name), Key = FCacheKey(Key), ValueIndex, Value = Value.RemoveData(), ValueIdGetter, OnCompletePtr = TRefCountPtr>(CompletionFunction)] (FHttpRequest::EResult HttpResult, FHttpRequest* Request) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_GetDataBatch_OnHttpRequestComplete); int64 ResponseCode = Request->GetResponseCode(); bool bHit = false; FCompressedBuffer CompressedBuffer; if (FHttpRequest::IsSuccessResponse(ResponseCode)) { FString ReceivedContentType; if (Request->GetHeader("Content-Type", ReceivedContentType)) { if (ReceivedContentType == TEXT("application/x-ue-comp")) { CompressedBuffer = FCompressedBuffer::FromCompressed(Request->MoveResponseBufferToShared()); bHit = true; } else if (ReceivedContentType == TEXT("application/octet-stream")) { CompressedBuffer = FValue::Compress(Request->MoveResponseBufferToShared()).GetData(); bHit = true; } else { bHit = false; } } else { CompressedBuffer = FCompressedBuffer::FromCompressed(Request->MoveResponseBufferToShared()); bHit = true; } } if (!ShouldAbortForShutdown() && !Owner.IsCanceled() && CacheStore.ShouldRetryOnError(HttpResult, ResponseCode) && ((Request->GetAttempts()+1) < UE_HTTPDDC_MAX_ATTEMPTS)) { return FHttpRequest::ECompletionBehavior::Retry; } if (!bHit) { UE_LOG(LogDerivedDataCache, Verbose, TEXT("%s: Cache miss with missing value %s with hash %s for %s from '%s'"), *CacheStore.GetName(), *ValueIdGetter(Value), *WriteToString<48>(Value.GetRawHash()), *WriteToString<96>(Key), *Name); OnCompletePtr->GetFunction()({ Name, Key, ValueIndex, Request->GetBytesReceived(), {}, EStatus::Error }); } else if (CompressedBuffer.GetRawHash() != Value.GetRawHash()) { UE_LOG(LogDerivedDataCache, Display, TEXT("%s: Cache miss with corrupted value %s with hash %s for %s from '%s'"), *CacheStore.GetName(), *ValueIdGetter(Value), *WriteToString<48>(Value.GetRawHash()), *WriteToString<96>(Key), *Name); OnCompletePtr->GetFunction()({ Name, Key, ValueIndex, Request->GetBytesReceived(), {}, EStatus::Error }); } else { OnCompletePtr->GetFunction()({ Name, Key, ValueIndex, Request->GetBytesReceived(), MoveTemp(CompressedBuffer), EStatus::Ok }); } return FHttpRequest::ECompletionBehavior::Done; }; TStringBuilder<256> CompressedBlobsUri; CompressedBlobsUri << "api/v1/compressed-blobs/" << CacheStore.StructuredNamespace << "/" << RawHash; Request->EnqueueAsyncDownload(Owner, Pool, *CompressedBlobsUri, MoveTemp(OnHttpRequestComplete), EHttpContentType::AnyContentType, { 404 }); } } FHttpCacheStore::FGetRecordOp::FGetRecordOp( FHttpCacheStore& InCacheStore, IRequestOwner& InOwner, const FSharedString& InName, const FCacheKey& InKey, uint64 InUserData, uint64 InBytesReceived, TArray&& InRequiredGets, TArray&& InRequiredHeads, FCacheRecordBuilder&& InRecordBuilder, TUniqueFunction&& InOnComplete) : CacheStore(InCacheStore) , Owner(InOwner) , Name(InName) , Key(InKey) , UserData(InUserData) , BytesReceived(InBytesReceived) , RequiredGets(MoveTemp(InRequiredGets)) , RequiredHeads(MoveTemp(InRequiredHeads)) , RecordBuilder(MoveTemp(InRecordBuilder)) , TotalOperations(RequiredGets.Num() + RequiredHeads.Num()) , SuccessfulOperations(0) , PendingOperations(TotalOperations) , OnComplete(MoveTemp(InOnComplete)) { FetchedBuffers.AddDefaulted(RequiredGets.Num()); } void FHttpCacheStore::FGetRecordOp::OnOnlyRecordComplete( FHttpCacheStore& CacheStore, IRequestOwner& Owner, const FCacheRecordPolicy& Policy, TUniqueFunction&& OnComplete, FGetCacheRecordOnlyResponse&& Response) { FCacheRecordBuilder RecordBuilder(Response.Key); if (Response.Status != EStatus::Ok) { return OnComplete({ Response.Name, RecordBuilder.Build(), Response.UserData, Response.Status }, Response.BytesReceived); } if (!EnumHasAnyFlags(Policy.GetRecordPolicy(), ECachePolicy::SkipMeta)) { RecordBuilder.SetMeta(FCbObject(Response.Record.Get().GetMeta())); } // TODO: There is not currently a batched GET endpoint for Jupiter. Once there is, all payload data should be fetched in one call. // In the meantime, we try to keep the code structured in a way that is friendly to future batching of GETs. TArray RequiredGets; TArray RequiredHeads; for (FValueWithId Value : Response.Record.Get().GetValues()) { const ECachePolicy ValuePolicy = Policy.GetValuePolicy(Value.GetId()); if (IsValueDataReady(Value, ValuePolicy)) { RecordBuilder.AddValue(MoveTemp(Value)); } else { if (EnumHasAnyFlags(ValuePolicy, ECachePolicy::SkipData)) { RequiredHeads.Emplace(Value); } else { RequiredGets.Emplace(Value); } } } if (RequiredGets.IsEmpty() && RequiredHeads.IsEmpty()) { return OnComplete({ Response.Name, RecordBuilder.Build(), Response.UserData, Response.Status }, Response.BytesReceived); } // Having this be a ref ensures we don't have the op reach 0 ref count in between the start of the exist batch operation and the get batch operation TRefCountPtr GetRecordOp = new FGetRecordOp( CacheStore, Owner, Response.Name, Response.Key, Response.UserData, Response.BytesReceived, MoveTemp(RequiredGets), MoveTemp(RequiredHeads), MoveTemp(RecordBuilder), MoveTemp(OnComplete) ); auto IdGetter = [](const FValueWithId& Value) { return FString(WriteToString<16>(Value.GetId())); }; { FRequestBarrier Barrier(Owner); GetRecordOp->DataProbablyExistsBatch(GetRecordOp->RequiredHeads, [GetRecordOp](FCachedDataProbablyExistsBatchResponse&& Response) { GetRecordOp->FinishDataStep(Response.Status == EStatus::Ok, 0); }); GetDataBatch(CacheStore, Owner, Response.Name, Response.Key, GetRecordOp->RequiredGets, IdGetter, [GetRecordOp](FGetCachedDataBatchResponse&& Response) { GetRecordOp->FetchedBuffers[Response.ValueIndex] = MoveTemp(Response.DataBuffer); GetRecordOp->FinishDataStep(Response.Status == EStatus::Ok, Response.BytesReceived); }); } } void FHttpCacheStore::FGetRecordOp::DataProbablyExistsBatch( TConstArrayView Values, FOnCachedDataProbablyExistsBatchComplete&& InOnComplete) { if (Values.IsEmpty()) { return; } FHttpRequestPool* Pool = nullptr; FHttpRequest* Request = CacheStore.WaitForHttpRequestForOwner(Owner, true /* bUnboundedOverflow */, Pool); TStringBuilder<256> CompressedBlobsUri; CompressedBlobsUri << "api/v1/compressed-blobs/" << CacheStore.StructuredNamespace << "/exists?"; bool bFirstItem = true; for (const FValueWithId& Value : Values) { if (!bFirstItem) { CompressedBlobsUri << "&"; } CompressedBlobsUri << "id=" << Value.GetRawHash(); bFirstItem = false; } auto OnHttpRequestComplete = [this, Values = TArray(Values), InOnComplete = MoveTemp(InOnComplete)](FHttpRequest::EResult HttpResult, FHttpRequest* Request) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_DataProbablyExistsBatch_OnHttpRequestComplete); int64 ResponseCode = Request->GetResponseCode(); if (FHttpRequest::IsSuccessResponse(ResponseCode)) { if (TSharedPtr ResponseObject = Request->GetResponseAsJsonObject()) { TArray NeedsArrayStrings; if (ResponseObject->TryGetStringArrayField(TEXT("needs"), NeedsArrayStrings)) { if (NeedsArrayStrings.IsEmpty()) { for (int32 ValueIndex = 0; ValueIndex < Values.Num(); ++ValueIndex) { const FValueWithId& Value = Values[ValueIndex]; UE_LOG(LogDerivedDataCache, VeryVerbose, TEXT("%s: Cache exists hit for %s with hash %s for %s from '%s'"), *CacheStore.GetName(), *WriteToString<16>(Value.GetId()), *WriteToString<48>(Value.GetRawHash()), *WriteToString<96>(Key), *Name); InOnComplete({ Name, Key, ValueIndex, EStatus::Ok }); } return FHttpRequest::ECompletionBehavior::Done; } } TBitArray<> ResultStatus(true, Values.Num()); for (const FString& NeedsString : NeedsArrayStrings) { const FIoHash NeedHash(NeedsString); for (int32 ValueIndex = 0; ValueIndex < Values.Num(); ++ValueIndex) { const FValueWithId& Value = Values[ValueIndex]; if (ResultStatus[ValueIndex] && NeedHash == Value.GetRawHash()) { ResultStatus[ValueIndex] = false; break; } } } for (int32 ValueIndex = 0; ValueIndex < Values.Num(); ++ValueIndex) { const FValueWithId& Value = Values[ValueIndex]; if (ResultStatus[ValueIndex]) { UE_LOG(LogDerivedDataCache, VeryVerbose, TEXT("%s: Cache exists hit for %s with hash %s for %s from '%s'"), *CacheStore.GetName(), *WriteToString<16>(Value.GetId()), *WriteToString<48>(Value.GetRawHash()), *WriteToString<96>(Key), *Name); InOnComplete({ Name, Key, ValueIndex, EStatus::Ok }); } else { UE_LOG(LogDerivedDataCache, Verbose, TEXT("%s: Cache exists miss with missing value %s with hash %s for %s from '%s'"), *CacheStore.GetName(), *WriteToString<16>(Value.GetId()), *WriteToString<48>(Value.GetRawHash()), *WriteToString<96>(Key), *Name); InOnComplete({ Name, Key, ValueIndex, EStatus::Error }); } } } else { for (int32 ValueIndex = 0; ValueIndex < Values.Num(); ++ValueIndex) { UE_LOG(LogDerivedDataCache, Log, TEXT("%s: Cache exists returned invalid results."), *CacheStore.GetName()); InOnComplete({ Name, Key, ValueIndex, EStatus::Error }); } } return FHttpRequest::ECompletionBehavior::Done; } if (!ShouldAbortForShutdown() && !Owner.IsCanceled() && CacheStore.ShouldRetryOnError(HttpResult, ResponseCode) && ((Request->GetAttempts()+1) < UE_HTTPDDC_MAX_ATTEMPTS)) { return FHttpRequest::ECompletionBehavior::Retry; } for (int32 ValueIndex = 0; ValueIndex < Values.Num(); ++ValueIndex) { const FValueWithId& Value = Values[ValueIndex]; UE_LOG(LogDerivedDataCache, Verbose, TEXT("%s: Cache miss with failed HTTP request for %s from '%s'"), *CacheStore.GetName(), *WriteToString<96>(Key), *Name); InOnComplete({Name, Key, ValueIndex, EStatus::Error}); } return FHttpRequest::ECompletionBehavior::Done; }; FCompositeBuffer DummyBuffer; Request->EnqueueAsyncPost(Owner, Pool, *CompressedBlobsUri, DummyBuffer, MoveTemp(OnHttpRequestComplete), EHttpContentType::FormUrlEncoded); } void FHttpCacheStore::FGetRecordOp::FinishDataStep(bool bSuccess, uint64 InBytesReceived) { BytesReceived.fetch_add(InBytesReceived, std::memory_order_relaxed); if (bSuccess) { SuccessfulOperations.fetch_add(1, std::memory_order_relaxed); } if (PendingOperations.fetch_sub(1, std::memory_order_acq_rel) == 1) { EStatus Status = EStatus::Error; uint32 LocalSuccessfulOperations = SuccessfulOperations.load(std::memory_order_relaxed); if (LocalSuccessfulOperations == TotalOperations) { for (int32 Index = 0; Index < RequiredHeads.Num(); ++Index) { RecordBuilder.AddValue(RequiredHeads[Index].RemoveData()); } for (int32 Index = 0; Index < RequiredGets.Num(); ++Index) { RecordBuilder.AddValue(FValueWithId(RequiredGets[Index].GetId(), FetchedBuffers[Index])); } Status = EStatus::Ok; } OnComplete({Name, RecordBuilder.Build(), UserData, Status}, BytesReceived.load(std::memory_order_relaxed)); } } FHttpCacheStore::FHttpCacheStore( const TCHAR* InServiceUrl, bool bResolveHostCanonicalName, const TCHAR* InNamespace, const TCHAR* InStructuredNamespace, const TCHAR* InOAuthProvider, const TCHAR* InOAuthClientId, const TCHAR* InOAuthSecret, const TCHAR* InOAuthScope, const EBackendLegacyMode InLegacyMode, const bool bInReadOnly) : Domain(InServiceUrl) , EffectiveDomain(InServiceUrl) , Namespace(InNamespace) , StructuredNamespace(InStructuredNamespace) , DefaultBucket(TEXT("default")) , OAuthProvider(InOAuthProvider) , OAuthClientId(InOAuthClientId) , OAuthSecret(InOAuthSecret) , OAuthScope(InOAuthScope) , Access(nullptr) , bIsUsable(false) , bReadOnly(bInReadOnly) , FailedLoginAttempts(0) , SpeedClass(ESpeedClass::Slow) , LegacyMode(InLegacyMode) { #if WITH_DATAREQUEST_HELPER FDataRequestHelper::StaticInitialize(); #endif SharedData = MakeUnique(); if (IsServiceReady() && AcquireAccessToken()) { FString OriginalDomainPrefix; TAnsiStringBuilder<64> DomainResolveName; if (Domain.StartsWith(TEXT("http://"))) { DomainResolveName << Domain.RightChop(7); OriginalDomainPrefix = TEXT("http://"); } else if (Domain.StartsWith(TEXT("https://"))) { DomainResolveName << Domain.RightChop(8); OriginalDomainPrefix = TEXT("https://"); } else { DomainResolveName << Domain; } addrinfo* AddrResult = nullptr; addrinfo AddrHints; FMemory::Memset(&AddrHints, 0, sizeof(AddrHints)); AddrHints.ai_flags = AI_CANONNAME; AddrHints.ai_family = AF_UNSPEC; if (bResolveHostCanonicalName && !::getaddrinfo(*DomainResolveName, nullptr, &AddrHints, &AddrResult)) { if (AddrResult->ai_canonname) { // Swap the domain with a canonical name from DNS so that if we are using regional redirection, we pin to a region. EffectiveDomain = OriginalDomainPrefix + ANSI_TO_TCHAR(AddrResult->ai_canonname); UE_LOG(LogDerivedDataCache, Display, TEXT("%s: Pinned to %s based on DNS canonical name."), *Domain, *EffectiveDomain); } else { EffectiveDomain = Domain; } ::freeaddrinfo(AddrResult); } else { EffectiveDomain = Domain; } GetRequestPools[0] = MakeUnique(*Domain, *EffectiveDomain, Access.Get(), SharedData.Get(), UE_HTTPDDC_GET_REQUEST_POOL_SIZE); GetRequestPools[1] = MakeUnique(*Domain, *EffectiveDomain, Access.Get(), SharedData.Get(), UE_HTTPDDC_GET_REQUEST_POOL_SIZE); PutRequestPools[0] = MakeUnique(*Domain, *EffectiveDomain, Access.Get(), SharedData.Get(), UE_HTTPDDC_PUT_REQUEST_POOL_SIZE); PutRequestPools[1] = MakeUnique(*Domain, *EffectiveDomain, Access.Get(), SharedData.Get(), UE_HTTPDDC_PUT_REQUEST_POOL_SIZE); // Allowing the non-blocking requests to overflow to double their pre-allocated size before we start waiting for one to free up. NonBlockingRequestPools = MakeUnique(*Domain, *EffectiveDomain, Access.Get(), SharedData.Get(), UE_HTTPDDC_NONBLOCKING_REQUEST_POOL_SIZE, UE_HTTPDDC_NONBLOCKING_REQUEST_POOL_SIZE); bIsUsable = true; } AnyInstance = this; } FHttpCacheStore::~FHttpCacheStore() { if (AnyInstance == this) { AnyInstance = nullptr; } #if WITH_DATAREQUEST_HELPER FDataRequestHelper::StaticShutdown(); #endif } FString FHttpCacheStore::GetName() const { return Domain; } TBitArray<> FHttpCacheStore::TryToPrefetch(TConstArrayView CacheKeys) { return CachedDataProbablyExistsBatch(CacheKeys); } bool FHttpCacheStore::WouldCache(const TCHAR* CacheKey, TArrayView InData) { return IsWritable(); } FHttpCacheStore::ESpeedClass FHttpCacheStore::GetSpeedClass() const { return SpeedClass; } bool FHttpCacheStore::ApplyDebugOptions(FBackendDebugOptions& InOptions) { DebugOptions = InOptions; return true; } bool FHttpCacheStore::IsServiceReady() { FHttpRequest Request(*Domain, *Domain, nullptr, SharedData.Get(), false); FHttpRequest::EResult Result = Request.PerformBlockingDownload(TEXT("health/ready"), nullptr); if (Result == FHttpRequest::EResult::Success && Request.GetResponseCode() == 200) { UE_LOG(LogDerivedDataCache, Display, TEXT("%s: HTTP DDC service status: %s."), *Request.GetName(), *Request.GetResponseAsString()); return true; } else { UE_LOG(LogDerivedDataCache, Warning, TEXT("%s: Unable to reach HTTP DDC service at %s. Status: %d . Response: %s"), *Request.GetName(), *Domain, Request.GetResponseCode(), *Request.GetResponseAsString()); } return false; } bool FHttpCacheStore::AcquireAccessToken() { if (Domain.StartsWith(TEXT("http://localhost"))) { UE_LOG(LogDerivedDataCache, Log, TEXT("Connecting to a local host '%s', so skipping authorization"), *Domain); return true; } // Avoid spamming the this if the service is down if (FailedLoginAttempts > UE_HTTPDDC_MAX_FAILED_LOGIN_ATTEMPTS) { return false; } ensureMsgf(OAuthProvider.StartsWith(TEXT("http://")) || OAuthProvider.StartsWith(TEXT("https://")), TEXT("The OAuth provider %s is not valid. Needs to be a fully qualified url."), *OAuthProvider ); // In case many requests wants to update the token at the same time // get the current serial while we wait to take the CS. const uint32 WantsToUpdateTokenSerial = Access.IsValid() ? Access->GetSerial() : 0u; { FScopeLock Lock(&AccessCs); // Check if someone has beaten us to update the token, then it // should now be valid. if (Access.IsValid() && Access->GetSerial() > WantsToUpdateTokenSerial) { return true; } const uint32 SchemeEnd = OAuthProvider.Find(TEXT("://")) + 3; const uint32 DomainEnd = OAuthProvider.Find(TEXT("/"), ESearchCase::CaseSensitive, ESearchDir::FromStart, SchemeEnd); FString AuthDomain(DomainEnd, *OAuthProvider); FString Uri(*OAuthProvider + DomainEnd + 1); FHttpRequest Request(*AuthDomain, *AuthDomain, nullptr, SharedData.Get(), false); FHttpRequest::EResult Result = FHttpRequest::EResult::Success; if (OAuthProvider.StartsWith(TEXT("http://localhost"))) { // Simple unauthenticated call to a local endpoint that mimics // the result from an OIDC provider. Result = Request.PerformBlockingDownload(*Uri, nullptr); } else { // Needs client id and secret to authenticate with an actual OIDC provider. // If contents of the secret string is a file path, resolve and read form data. if (OAuthSecret.StartsWith(TEXT("file://"))) { FString FilePath = OAuthSecret.Mid(7, OAuthSecret.Len() - 7); FString SecretFileContents; if (FFileHelper::LoadFileToString(SecretFileContents, *FilePath)) { // Overwrite the filepath with the actual content. OAuthSecret = SecretFileContents; } else { UE_LOG(LogDerivedDataCache, Warning, TEXT("%s: Failed to read OAuth form data file (%s)."), *Request.GetName(), *OAuthSecret); return false; } } FString OAuthFormData = FString::Printf( TEXT("client_id=%s&scope=%s&grant_type=client_credentials&client_secret=%s"), *OAuthClientId, *OAuthScope, *OAuthSecret ); TArray64 FormData; auto OAuthFormDataUTF8 = FTCHARToUTF8(*OAuthFormData); FormData.Append((uint8*)OAuthFormDataUTF8.Get(), OAuthFormDataUTF8.Length()); Result = Request.PerformBlockingPost(*Uri, FCompositeBuffer(FSharedBuffer::MakeView(FormData.GetData(), FormData.Num())), EHttpContentType::FormUrlEncoded); } if (Result == FHttpRequest::EResult::Success && Request.GetResponseCode() == 200) { TSharedPtr ResponseObject = Request.GetResponseAsJsonObject(); if (ResponseObject) { FString AccessTokenString; int32 ExpiryTimeSeconds = 0; int32 CurrentTimeSeconds = int32(FPlatformTime::ToSeconds(FPlatformTime::Cycles())); if (ResponseObject->TryGetStringField(TEXT("access_token"), AccessTokenString) && ResponseObject->TryGetNumberField(TEXT("expires_in"), ExpiryTimeSeconds)) { if (!Access) { Access = MakeUnique(); } Access->SetHeader(*AccessTokenString); UE_LOG(LogDerivedDataCache, Display, TEXT("%s: Logged in to HTTP DDC services. Expires in %d seconds."), *Request.GetName(), ExpiryTimeSeconds); //Schedule a refresh of the token ahead of expiry time (this will not work in commandlets) if (!IsRunningCommandlet()) { FTSTicker::GetCoreTicker().AddTicker(FTickerDelegate::CreateLambda( [this](float DeltaTime) { this->AcquireAccessToken(); return false; } ), ExpiryTimeSeconds - 20.0f); } // Reset failed login attempts, the service is indeed alive. FailedLoginAttempts = 0; return true; } } } else { UE_LOG(LogDerivedDataCache, Warning, TEXT("%s: Failed to log in to HTTP services. Server responed with code %d."), *Request.GetName(), Request.GetResponseCode()); FailedLoginAttempts++; } } return false; } bool FHttpCacheStore::ShouldRetryOnError(FHttpRequest::EResult Result, int64 ResponseCode) { if (Result == FHttpRequest::EResult::FailedTimeout) { return true; } // Access token might have expired, request a new token and try again. if (ResponseCode == 401 && AcquireAccessToken()) { return true; } // Too many requests, make a new attempt if (ResponseCode == 429) { return true; } return false; } void FHttpCacheStore::GetCacheRecordOnlyAsync( IRequestOwner& Owner, const FSharedString& Name, const FCacheKey& Key, const FCacheRecordPolicy& Policy, uint64 UserData, FOnGetCacheRecordOnlyComplete&& OnComplete) { auto MakeResponse = [Name = FSharedString(Name), Key, UserData](uint64 BytesReceived, EStatus Status) { return FGetCacheRecordOnlyResponse{ Name, Key, UserData, BytesReceived, {}, Status }; }; if (!IsUsable()) { UE_LOG(LogDerivedDataCache, VeryVerbose, TEXT("%s: Skipped get of %s from '%s' because this cache store is not available"), *GetName(), *WriteToString<96>(Key), *Name); return OnComplete(MakeResponse(0, EStatus::Error)); } // Skip the request if querying the cache is disabled. const ECachePolicy QueryPolicy = SpeedClass == ESpeedClass::Local ? ECachePolicy::QueryLocal : ECachePolicy::QueryRemote; if (!EnumHasAnyFlags(Policy.GetRecordPolicy(), QueryPolicy)) { UE_LOG(LogDerivedDataCache, VeryVerbose, TEXT("%s: Skipped get of %s from '%s' due to cache policy"), *GetName(), *WriteToString<96>(Key), *Name); return OnComplete(MakeResponse(0, EStatus::Error)); } if (DebugOptions.ShouldSimulateGetMiss(Key)) { UE_LOG(LogDerivedDataCache, Verbose, TEXT("%s: Simulated miss for get of %s from '%s'"), *GetName(), *WriteToString<96>(Key), *Name); return OnComplete(MakeResponse(0, EStatus::Error)); } FString Bucket(Key.Bucket.ToString()); Bucket.ToLowerInline(); TStringBuilder<256> RefsUri; RefsUri << "api/v1/refs/" << StructuredNamespace << "/" << Bucket << "/" << Key.Hash; FHttpRequestPool* Pool = nullptr; FHttpRequest* Request = WaitForHttpRequestForOwner(Owner, false /* bUnboundedOverflow */, Pool); auto OnHttpRequestComplete = [this, &Owner, Name = FSharedString(Name), Key, UserData, OnComplete = MoveTemp(OnComplete)] (FHttpRequest::EResult HttpResult, FHttpRequest* Request) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_GetCacheRecordOnlyAsync_OnHttpRequestComplete); int64 ResponseCode = Request->GetResponseCode(); if (FHttpRequest::IsSuccessResponse(ResponseCode)) { FSharedBuffer ResponseBuffer = Request->MoveResponseBufferToShared(); if (ValidateCompactBinary(ResponseBuffer, ECbValidateMode::Default) != ECbValidateError::None) { UE_LOG(LogDerivedDataCache, Log, TEXT("%s: Cache miss with invalid package for %s from '%s'"), *GetName(), *WriteToString<96>(Key), *Name); OnComplete({ Name, Key, UserData, Request->GetBytesReceived(), {}, EStatus::Error }); return FHttpRequest::ECompletionBehavior::Done; } FOptionalCacheRecord Record = FCacheRecord::Load(FCbPackage(FCbObject(ResponseBuffer))); if (Record.IsNull()) { UE_LOG(LogDerivedDataCache, Log, TEXT("%s: Cache miss with record load failure for %s from '%s'"), *GetName(), *WriteToString<96>(Key), *Name); OnComplete({ Name, Key, UserData, Request->GetBytesReceived(), {}, EStatus::Error }); return FHttpRequest::ECompletionBehavior::Done; } OnComplete({ Name, Key, UserData, Request->GetBytesReceived(), MoveTemp(Record), EStatus::Ok }); return FHttpRequest::ECompletionBehavior::Done; } if (!ShouldAbortForShutdown() && !Owner.IsCanceled() && ShouldRetryOnError(HttpResult, ResponseCode) && ((Request->GetAttempts()+1) < UE_HTTPDDC_MAX_ATTEMPTS)) { return FHttpRequest::ECompletionBehavior::Retry; } UE_LOG(LogDerivedDataCache, Verbose, TEXT("%s: Cache miss with missing package for %s from '%s'"), *GetName(), *WriteToString<96>(Key), *Name); OnComplete({ Name, Key, UserData, Request->GetBytesReceived(), {}, EStatus::Error }); return FHttpRequest::ECompletionBehavior::Done; }; Request->EnqueueAsyncDownload(Owner, Pool, *RefsUri, MoveTemp(OnHttpRequestComplete), EHttpContentType::CbObject, { 401, 404 }); } void FHttpCacheStore::PutCacheRecordAsync( IRequestOwner& Owner, const FSharedString& Name, const FCacheRecord& Record, const FCacheRecordPolicy& Policy, uint64 UserData, TUniqueFunction&& OnComplete) { const FCacheKey& Key = Record.GetKey(); auto MakeResponse = [Name = FSharedString(Name), Key = FCacheKey(Key), UserData](EStatus Status) { return FCachePutResponse{ Name, Key, UserData, Status }; }; if (!IsWritable()) { UE_LOG(LogDerivedDataCache, VeryVerbose, TEXT("%s: Skipped put of %s from '%s' because this cache store is read-only"), *GetName(), *WriteToString<96>(Key), *Name); return OnComplete(MakeResponse(EStatus::Error), 0); } // Skip the request if storing to the cache is disabled. const ECachePolicy RecordPolicy = Policy.GetRecordPolicy(); const ECachePolicy StoreFlag = SpeedClass == ESpeedClass::Local ? ECachePolicy::StoreLocal : ECachePolicy::StoreRemote; if (!EnumHasAnyFlags(RecordPolicy, StoreFlag)) { UE_LOG(LogDerivedDataCache, VeryVerbose, TEXT("%s: Skipped put of %s from '%s' due to cache policy"), *GetName(), *WriteToString<96>(Key), *Name); return OnComplete(MakeResponse(EStatus::Error), 0); } if (DebugOptions.ShouldSimulatePutMiss(Key)) { UE_LOG(LogDerivedDataCache, Verbose, TEXT("%s: Simulated miss for put of %s from '%s'"), *GetName(), *WriteToString<96>(Key), *Name); return OnComplete(MakeResponse(EStatus::Error), 0); } // TODO: Jupiter currently always overwrites. It doesn't have a "write if not present" feature (for records or attachments), // but would require one to implement all policy correctly. FString Bucket(Key.Bucket.ToString()); Bucket.ToLowerInline(); FCbPackage Package = Record.Save(); FPutPackageOp::PutPackage(*this, Owner, Name, Key, MoveTemp(Package), Policy, UserData, [MakeResponse = MoveTemp(MakeResponse), OnComplete = MoveTemp(OnComplete)](FPutPackageOp::FCachePutPackageResponse&& Response) { OnComplete(MakeResponse(Response.Status), Response.BytesSent); }); } void FHttpCacheStore::PutCacheValueAsync( IRequestOwner& Owner, const FSharedString& Name, const FCacheKey& Key, const FValue& Value, const ECachePolicy Policy, uint64 UserData, TUniqueFunction&& OnComplete) { auto MakeResponse = [Name = FSharedString(Name), Key = FCacheKey(Key), UserData](EStatus Status) { return FCachePutValueResponse{ Name, Key, UserData, Status }; }; if (!IsWritable()) { UE_LOG(LogDerivedDataCache, VeryVerbose, TEXT("%s: Skipped put of %s from '%s' because this cache store is read-only"), *GetName(), *WriteToString<96>(Key), *Name); return OnComplete(MakeResponse(EStatus::Error), 0); } // Skip the request if storing to the cache is disabled. const ECachePolicy StoreFlag = SpeedClass == ESpeedClass::Local ? ECachePolicy::StoreLocal : ECachePolicy::StoreRemote; if (!EnumHasAnyFlags(Policy, StoreFlag)) { UE_LOG(LogDerivedDataCache, VeryVerbose, TEXT("%s: Skipped put of %s from '%s' due to cache policy"), *GetName(), *WriteToString<96>(Key), *Name); return OnComplete(MakeResponse(EStatus::Error), 0); } if (DebugOptions.ShouldSimulatePutMiss(Key)) { UE_LOG(LogDerivedDataCache, Verbose, TEXT("%s: Simulated miss for put of %s from '%s'"), *GetName(), *WriteToString<96>(Key), *Name); return OnComplete(MakeResponse(EStatus::Error), 0); } // TODO: Jupiter currently always overwrites. It doesn't have a "write if not present" feature (for records or attachments), // but would require one to implement all policy correctly. FString Bucket(Key.Bucket.ToString()); Bucket.ToLowerInline(); FCbWriter Writer; Writer.BeginObject(); Writer.AddBinaryAttachment("RawHash", Value.GetRawHash()); Writer.AddInteger("RawSize", Value.GetRawSize()); Writer.EndObject(); FCbPackage Package(Writer.Save().AsObject()); Package.AddAttachment(FCbAttachment(Value.GetData())); FPutPackageOp::PutPackage(*this, Owner, Name, Key, MoveTemp(Package), Policy, UserData, [MakeResponse = MoveTemp(MakeResponse), OnComplete = MoveTemp(OnComplete)](FPutPackageOp::FCachePutPackageResponse&& Response) { OnComplete(MakeResponse(Response.Status), Response.BytesSent); }); } void FHttpCacheStore::GetCacheValueAsync( IRequestOwner& Owner, FSharedString Name, const FCacheKey& Key, ECachePolicy Policy, uint64 UserData, FOnCacheGetValueComplete&& OnComplete) { if (!IsUsable()) { UE_LOG(LogDerivedDataCache, VeryVerbose, TEXT("%s: Skipped get of %s from '%s' because this cache store is not available"), *GetName(), *WriteToString<96>(Key), *Name); OnComplete({Name, Key, {}, UserData, EStatus::Error}); return; } // Skip the request if querying the cache is disabled. const ECachePolicy QueryFlag = SpeedClass == ESpeedClass::Local ? ECachePolicy::QueryLocal : ECachePolicy::QueryRemote; if (!EnumHasAnyFlags(Policy, QueryFlag)) { UE_LOG(LogDerivedDataCache, VeryVerbose, TEXT("%s: Skipped get of %s from '%s' due to cache policy"), *GetName(), *WriteToString<96>(Key), *Name); OnComplete({Name, Key, {}, UserData, EStatus::Error}); return; } if (DebugOptions.ShouldSimulateGetMiss(Key)) { UE_LOG(LogDerivedDataCache, Verbose, TEXT("%s: Simulated miss for get of %s from '%s'"), *GetName(), *WriteToString<96>(Key), *Name); OnComplete({Name, Key, {}, UserData, EStatus::Error}); return; } const bool bSkipData = EnumHasAnyFlags(Policy, ECachePolicy::SkipData); FString Bucket(Key.Bucket.ToString()); Bucket.ToLowerInline(); TStringBuilder<256> RefsUri; RefsUri << "api/v1/refs/" << StructuredNamespace << "/" << Bucket << "/" << Key.Hash; FHttpRequestPool* Pool = nullptr; FHttpRequest* Request = WaitForHttpRequestForOwner(Owner, false /* bUnboundedOverflow */, Pool); if (bSkipData) { Request->AddHeader(TEXT("Accept"), TEXT("application/x-ue-cb")); } else { Request->AddHeader(TEXT("Accept"), TEXT("application/x-jupiter-inline")); } auto OnHttpRequestComplete = [this, &Owner, Name, Key, UserData, bSkipData, OnComplete = MoveTemp(OnComplete)] (FHttpRequest::EResult HttpResult, FHttpRequest* Request) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_GetCacheValueAsync_OnHttpRequestComplete); int64 ResponseCode = Request->GetResponseCode(); if (FHttpRequest::IsSuccessResponse(ResponseCode)) { FValue ResultValue; FSharedBuffer ResponseBuffer = Request->MoveResponseBufferToShared(); if (bSkipData) { if (ValidateCompactBinary(ResponseBuffer, ECbValidateMode::Default) != ECbValidateError::None) { UE_LOG(LogDerivedDataCache, Display, TEXT("%s: Cache miss with invalid package for %s from '%s'"), *GetName(), *WriteToString<96>(Key), *Name); OnComplete({Name, Key, {}, UserData, EStatus::Error}); return FHttpRequest::ECompletionBehavior::Done; } const FCbObjectView Object = FCbObject(ResponseBuffer); const FIoHash RawHash = Object["RawHash"].AsHash(); const uint64 RawSize = Object["RawSize"].AsUInt64(MAX_uint64); if (RawHash.IsZero() || RawSize == MAX_uint64) { UE_LOG(LogDerivedDataCache, Display, TEXT("%s: Cache miss with invalid value for %s from '%'"), *GetName(), *WriteToString<96>(Key), *Name); OnComplete({Name, Key, {}, UserData, EStatus::Error}); return FHttpRequest::ECompletionBehavior::Done; } ResultValue = FValue(RawHash, RawSize); } else { FCompressedBuffer CompressedBuffer = FCompressedBuffer::FromCompressed(ResponseBuffer); if (!CompressedBuffer) { FString ReceivedHashStr; if (Request->GetHeader("X-Jupiter-InlinePayloadHash", ReceivedHashStr)) { FIoHash ReceivedHash(ReceivedHashStr); FIoHash ComputedHash = FIoHash::HashBuffer(ResponseBuffer.GetView()); if (ReceivedHash == ComputedHash) { CompressedBuffer = FCompressedBuffer::Compress(ResponseBuffer); } } } if (!CompressedBuffer) { UE_LOG(LogDerivedDataCache, Display, TEXT("%s: Cache miss with invalid package for %s from '%s'"), *GetName(), *WriteToString<96>(Key), *Name); OnComplete({Name, Key, {}, UserData, EStatus::Error}); return FHttpRequest::ECompletionBehavior::Done; } ResultValue = FValue(CompressedBuffer); } OnComplete({Name, Key, ResultValue, UserData, EStatus::Ok}); return FHttpRequest::ECompletionBehavior::Done; } if (!ShouldAbortForShutdown() && !Owner.IsCanceled() && ShouldRetryOnError(HttpResult, ResponseCode) && ((Request->GetAttempts()+1) < UE_HTTPDDC_MAX_ATTEMPTS)) { return FHttpRequest::ECompletionBehavior::Retry; } UE_LOG(LogDerivedDataCache, Verbose, TEXT("%s: Cache miss with failed HTTP request for %s from '%s'"), *GetName(), *WriteToString<96>(Key), *Name); OnComplete({Name, Key, {}, UserData, EStatus::Error}); return FHttpRequest::ECompletionBehavior::Done; }; Request->EnqueueAsyncDownload(Owner, Pool, *RefsUri, MoveTemp(OnHttpRequestComplete), EHttpContentType::UnspecifiedContentType, { 401, 404 }); } void FHttpCacheStore::GetCacheRecordAsync( IRequestOwner& Owner, const FSharedString& Name, const FCacheKey& Key, const FCacheRecordPolicy& Policy, uint64 UserData, TUniqueFunction&& OnComplete) { FGetRecordOp::GetRecord(*this, Owner, Name, Key, Policy, UserData, MoveTemp(OnComplete)); } void FHttpCacheStore::RefCachedDataProbablyExistsBatchAsync( IRequestOwner& Owner, TConstArrayView ValueRefs, FOnCacheGetValueComplete&& OnComplete) { if (ValueRefs.IsEmpty()) { return; } if (!IsUsable()) { for (const FCacheGetValueRequest& ValueRef : ValueRefs) { UE_LOG(LogDerivedDataCache, VeryVerbose, TEXT("%s: Skipped exists check of %s from '%s' because this cache store is not available"), *GetName(), *WriteToString<96>(ValueRef.Key), *ValueRef.Name); OnComplete(ValueRef.MakeResponse(EStatus::Error)); } return; } TStringBuilder<256> RefsUri; RefsUri << "api/v1/refs/" << StructuredNamespace; FCbWriter RequestWriter; RequestWriter.BeginObject(); RequestWriter.BeginArray(ANSITEXTVIEW("ops")); uint32 OpIndex = 0; for (const FCacheGetValueRequest& ValueRef : ValueRefs) { RequestWriter.BeginObject(); RequestWriter.AddInteger(ANSITEXTVIEW("opId"), OpIndex); RequestWriter.AddString(ANSITEXTVIEW("op"), ANSITEXTVIEW("GET")); FCacheKey Key = ValueRef.Key; FString Bucket(Key.Bucket.ToString()); Bucket.ToLowerInline(); RequestWriter.AddString(ANSITEXTVIEW("bucket"), Bucket); RequestWriter.AddString(ANSITEXTVIEW("key"), LexToString(Key.Hash)); RequestWriter.AddBool(ANSITEXTVIEW("resolveAttachments"), true); RequestWriter.EndObject(); ++OpIndex; } RequestWriter.EndArray(); RequestWriter.EndObject(); FCbFieldIterator RequestFields = RequestWriter.Save(); FHttpRequestPool* Pool = nullptr; FHttpRequest* Request = WaitForHttpRequestForOwner(Owner, false /* bUnboundedOverflow */, Pool); Request->AddHeader(TEXT("Accept"), TEXT("application/x-ue-cb")); auto OnHttpRequestComplete = [this, &Owner, ValueRefs = TArray(ValueRefs), OnComplete = MoveTemp(OnComplete)](FHttpRequest::EResult HttpResult, FHttpRequest* Request) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_RefCachedDataProbablyExistsBatchAsync_OnHttpRequestComplete); int64 ResponseCode = Request->GetResponseCode(); if (FHttpRequest::IsSuccessResponse(ResponseCode)) { FMemoryView ResponseView = MakeMemoryView(Request->GetResponseBuffer().GetData(), Request->GetResponseBuffer().Num()); if (ValidateCompactBinary(ResponseView, ECbValidateMode::Default) != ECbValidateError::None) { for (const FCacheGetValueRequest& ValueRef : ValueRefs) { UE_LOG(LogDerivedDataCache, Log, TEXT("%s: Cache exists returned invalid results."), *GetName()); OnComplete(ValueRef.MakeResponse(EStatus::Error)); } return FHttpRequest::ECompletionBehavior::Done; } const FCbObjectView ResponseObject = FCbObjectView(Request->GetResponseBuffer().GetData()); FCbArrayView ResultsArrayView = ResponseObject[ANSITEXTVIEW("results")].AsArrayView(); if (ResultsArrayView.Num() != ValueRefs.Num()) { for (const FCacheGetValueRequest& ValueRef : ValueRefs) { UE_LOG(LogDerivedDataCache, Log, TEXT("%s: Cache exists returned unexpected quantity of results (expected %d, got %d)."), *GetName(), ValueRefs.Num(), ResultsArrayView.Num()); OnComplete(ValueRef.MakeResponse(EStatus::Error)); } return FHttpRequest::ECompletionBehavior::Done; } for (FCbFieldView ResultFieldView : ResultsArrayView) { FCbObjectView ResultObjectView = ResultFieldView.AsObjectView(); uint32 OpId = ResultObjectView[ANSITEXTVIEW("opId")].AsUInt32(); FCbObjectView ResponseObjectView = ResultObjectView[ANSITEXTVIEW("response")].AsObjectView(); int32 StatusCode = ResultObjectView[ANSITEXTVIEW("statusCode")].AsInt32(); if (OpId >= (uint32)ValueRefs.Num()) { UE_LOG(LogDerivedDataCache, Display, TEXT("%s: Encountered invalid opId %d while querying %d values"), *GetName(), OpId, ValueRefs.Num()); continue; } const FCacheGetValueRequest& ValueRef = ValueRefs[OpId]; if (!FHttpRequest::IsSuccessResponse(StatusCode)) { UE_LOG(LogDerivedDataCache, Verbose, TEXT("%s: Cache miss with unsuccessful response code %d for %s from '%s'"), *GetName(), StatusCode, *WriteToString<96>(ValueRef.Key), *ValueRef.Name); OnComplete(ValueRef.MakeResponse(EStatus::Error)); continue; } const ECachePolicy QueryFlag = SpeedClass == ESpeedClass::Local ? ECachePolicy::QueryLocal : ECachePolicy::QueryRemote; if (!EnumHasAnyFlags(ValueRef.Policy, QueryFlag)) { UE_LOG(LogDerivedDataCache, VeryVerbose, TEXT("%s: Skipped exists check of %s from '%s' due to cache policy"), *GetName(), *WriteToString<96>(ValueRef.Key), *ValueRef.Name); OnComplete(ValueRef.MakeResponse(EStatus::Error)); continue; } const FIoHash RawHash = ResponseObjectView[ANSITEXTVIEW("RawHash")].AsHash(); const uint64 RawSize = ResponseObjectView[ANSITEXTVIEW("RawSize")].AsUInt64(MAX_uint64); if (RawHash.IsZero() || RawSize == MAX_uint64) { UE_LOG(LogDerivedDataCache, Display, TEXT("%s: Cache miss with invalid value for %s from '%s'"), *GetName(), *WriteToString<96>(ValueRef.Key), *ValueRef.Name); OnComplete(ValueRef.MakeResponse(EStatus::Error)); continue; } OnComplete({ValueRef.Name, ValueRef.Key, FValue(RawHash, RawSize), ValueRef.UserData, EStatus::Ok}); } return FHttpRequest::ECompletionBehavior::Done; } if (!ShouldAbortForShutdown() && !Owner.IsCanceled() && ShouldRetryOnError(HttpResult, ResponseCode) && ((Request->GetAttempts()+1) < UE_HTTPDDC_MAX_ATTEMPTS)) { return FHttpRequest::ECompletionBehavior::Retry; } for (const FCacheGetValueRequest& ValueRef : ValueRefs) { UE_LOG(LogDerivedDataCache, Verbose, TEXT("%s: Cache miss with failed HTTP request for %s from '%s'"), *GetName(), *WriteToString<96>(ValueRef.Key), *ValueRef.Name); OnComplete(ValueRef.MakeResponse(EStatus::Error)); } return FHttpRequest::ECompletionBehavior::Done; }; Request->EnqueueAsyncPost(Owner, Pool, *RefsUri, FCompositeBuffer(RequestFields.GetOuterBuffer()), MoveTemp(OnHttpRequestComplete), EHttpContentType::CbObject); } bool FHttpCacheStore::CachedDataProbablyExists(const TCHAR* CacheKey) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_Exist); TRACE_COUNTER_ADD(HttpDDC_Exist, int64(1)); COOK_STAT(auto Timer = UsageStats.TimeProbablyExists()); if (DebugOptions.ShouldSimulateGetMiss(CacheKey)) { return false; } #if WITH_DATAREQUEST_HELPER // Retry request until we get an accepted response or exhaust allowed number of attempts. for (int32 Attempts = 0; Attempts < UE_HTTPDDC_MAX_ATTEMPTS; ++Attempts) { FDataRequestHelper RequestHelper(GetRequestPools[IsInGameThread()].Get(), *Namespace, *DefaultBucket, CacheKey, nullptr); const int64 ResponseCode = RequestHelper.GetResponseCode(); if (FHttpRequest::IsSuccessResponse(ResponseCode) && RequestHelper.IsSuccess()) { COOK_STAT(Timer.AddHit(0)); return true; } if (!ShouldRetryOnError(ResponseCode)) { return false; } } #else FString Uri = FString::Printf(TEXT("api/v1/c/ddc/%s/%s/%s"), *Namespace, *DefaultBucket, CacheKey); // Retry request until we get an accepted response or exhaust allowed number of attempts. for (int32 Attempts = 0; Attempts < UE_HTTPDDC_MAX_ATTEMPTS; ++Attempts) { FScopedHttpPoolRequestPtr Request(GetRequestPools[IsInGameThread()].Get()); const FHttpRequest::EResult Result = Request->PerformBlockingQuery(*Uri); const int64 ResponseCode = Request->GetResponseCode(); if (FHttpRequest::IsSuccessResponse(ResponseCode) || ResponseCode == 400) { const bool bIsHit = (Result == FHttpRequest::Success && FHttpRequest::IsSuccessResponse(ResponseCode)); if (bIsHit) { TRACE_COUNTER_ADD(HttpDDC_ExistHit, int64(1)); COOK_STAT(Timer.AddHit(0)); } return bIsHit; } if (!ShouldRetryOnError(ResponseCode)) { break; } } #endif return false; } TBitArray<> FHttpCacheStore::CachedDataProbablyExistsBatch(TConstArrayView CacheKeys) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_Exist); TRACE_COUNTER_ADD(HttpDDC_Exist, int64(1)); COOK_STAT(auto Timer = UsageStats.TimeProbablyExists()); #if WITH_DATAREQUEST_HELPER for (int32 Attempts = 0; Attempts < UE_HTTPDDC_MAX_ATTEMPTS; ++Attempts) { FDataRequestHelper RequestHelper(GetRequestPools[IsInGameThread()].Get(), *Namespace, *DefaultBucket, CacheKeys); const int64 ResponseCode = RequestHelper.GetResponseCode(); if (FHttpRequest::IsSuccessResponse(ResponseCode) && RequestHelper.IsSuccess()) { COOK_STAT(Timer.AddHit(0)); TBitArray<> Results = RequestHelper.IsBatchSuccess(); int32 ResultIndex = 0; for (const FString& CacheKey : CacheKeys) { if (DebugOptions.ShouldSimulateGetMiss(*CacheKey)) { Results[ResultIndex] = false; } ResultIndex++; } return Results; } if (!ShouldRetryOnError(ResponseCode)) { TBitArray<> Results = RequestHelper.IsBatchSuccess(); int32 ResultIndex = 0; for (const FString& CacheKey : CacheKeys) { if (DebugOptions.ShouldSimulateGetMiss(*CacheKey)) { Results[ResultIndex] = false; } ResultIndex++; } return Results; } } #else const TCHAR* const Uri = TEXT("api/v1/c/ddc-rpc"); TAnsiStringBuilder<512> Body; const FTCHARToUTF8 AnsiNamespace(*Namespace); const FTCHARToUTF8 AnsiBucket(*DefaultBucket); Body << "{\"Operations\":["; for (const FString& CacheKey : CacheKeys) { Body << "{\"Namespace\":\"" << AnsiNamespace.Get() << "\",\"Bucket\":\"" << AnsiBucket.Get() << "\","; Body << "\"Id\":\"" << FTCHARToUTF8(*CacheKey).Get() << "\",\"Op\":\"HEAD\"},"; } Body.RemoveSuffix(1); Body << "]}"; TConstArrayView BodyView(reinterpret_cast(Body.ToString()), Body.Len()); // Retry request until we get an accepted response or exhaust allowed number of attempts. for (int32 Attempts = 0; Attempts < UE_HTTPDDC_MAX_ATTEMPTS; ++Attempts) { FScopedHttpPoolRequestPtr Request(GetRequestPools[IsInGameThread()].Get()); const FHttpRequest::EResult Result = Request->PerformBlockingUpload(Uri, BodyView); const int64 ResponseCode = Request->GetResponseCode(); if (Result == FHttpRequest::Success && ResponseCode == 200) { TArray> ResponseArray = Request->GetResponseAsJsonArray(); TBitArray<> Exists; Exists.Reserve(CacheKeys.Num()); for (const FString& CacheKey : CacheKeys) { if (DebugOptions.ShouldSimulateGetMiss(*CacheKey)) { Exists.Add(false); } else { const TSharedPtr* FoundResponse = Algo::FindByPredicate(ResponseArray, [&CacheKey](const TSharedPtr& Response) { FString Key; Response->TryGetString(Key); return Key == CacheKey; }); Exists.Add(FoundResponse != nullptr); } } if (Exists.CountSetBits() == CacheKeys.Num()) { TRACE_COUNTER_ADD(HttpDDC_ExistHit, int64(1)); COOK_STAT(Timer.AddHit(0)); } return Exists; } if (!ShouldRetryOnError(ResponseCode)) { break; } } #endif return TBitArray<>(false, CacheKeys.Num()); } bool FHttpCacheStore::GetCachedData(const TCHAR* CacheKey, TArray& OutData) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_GetCachedData); TRACE_COUNTER_ADD(HttpDDC_Get, int64(1)); COOK_STAT(auto Timer = UsageStats.TimeGet()); if (DebugOptions.ShouldSimulateGetMiss(CacheKey)) { return false; } TArray64 ArrayBuffer; #if WITH_DATAREQUEST_HELPER // Retry request until we get an accepted response or exhaust allowed number of attempts. for (int32 Attempts = 0; Attempts < UE_HTTPDDC_MAX_ATTEMPTS; ++Attempts) { FDataRequestHelper RequestHelper(GetRequestPools[IsInGameThread()].Get(), *Namespace, *DefaultBucket, CacheKey, &ArrayBuffer); OutData = TArray(MoveTemp(ArrayBuffer)); const int64 ResponseCode = RequestHelper.GetResponseCode(); if (FHttpRequest::IsSuccessResponse(ResponseCode) && RequestHelper.IsSuccess()) { COOK_STAT(Timer.AddHit(OutData.Num())); check(OutData.Num() > 0); return true; } if (!ShouldRetryOnError(ResponseCode)) { return false; } } #else FString Uri = FString::Printf(TEXT("api/v1/c/ddc/%s/%s/%s.raw"), *Namespace, *DefaultBucket, CacheKey); // Retry request until we get an accepted response or exhaust allowed number of attempts. for (int32 Attempts = 0; Attempts < UE_HTTPDDC_MAX_ATTEMPTS; ++Attempts) { FScopedHttpPoolRequestPtr Request(GetRequestPools[IsInGameThread()].Get()); if (Request.IsValid()) { FHttpRequest::EResult Result = Request->PerformBlockingDownload(*Uri, &ArrayBuffer); const uint64 ResponseCode = Request->GetResponseCode(); // Request was successful, make sure we got all the expected data. if (FHttpRequest::IsSuccessResponse(ResponseCode) && VerifyRequest(Request.Get(), *Namespace, *DefaultBucket, CacheKey, ArrayBuffer)) { OutData = TArray(MoveTemp(ArrayBuffer)); TRACE_COUNTER_ADD(HttpDDC_GetHit, int64(1)); TRACE_COUNTER_ADD(HttpDDC_BytesReceived, int64(Request->GetBytesReceived())); COOK_STAT(Timer.AddHit(Request->GetBytesReceived())); return true; } OutData = TArray(MoveTemp(ArrayBuffer)); if (!ShouldRetryOnError(ResponseCode)) { return false; } } } #endif return false; } FDerivedDataBackendInterface::EPutStatus FHttpCacheStore::PutCachedData(const TCHAR* CacheKey, TArrayView InData, bool bPutEvenIfExists) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_PutCachedData); if (!IsWritable()) { UE_LOG(LogDerivedDataCache, Verbose, TEXT("%s is read only. Skipping put of %s"), *GetName(), CacheKey); return EPutStatus::NotCached; } // don't put anything we pretended didn't exist if (DebugOptions.ShouldSimulatePutMiss(CacheKey)) { return EPutStatus::Skipped; } #if 0 // No longer WITH_DATAREQUEST_HELPER as async puts are unsupported except through the AsyncPutWrapper which expects the inner backend to perform the put synchronously for (int32 Attempts = 0; Attempts < UE_HTTPDDC_MAX_ATTEMPTS; ++Attempts) { FDataUploadHelper Request(PutRequestPools[IsInGameThread()].Get(), *Namespace, *DefaultBucket, CacheKey, InData, UsageStats); if (ShouldAbortForShutdown()) { return EPutStatus::NotCached; } const int64 ResponseCode = Request.GetResponseCode(); if (Request.IsSuccess() && (Request.IsQueued() || FHttpRequest::IsSuccessResponse(ResponseCode))) { return Request.IsQueued() ? EPutStatus::Executing : EPutStatus::Cached; } if (!ShouldRetryOnError(ResponseCode)) { return EPutStatus::NotCached; } } #else COOK_STAT(auto Timer = UsageStats.TimePut()); FString Uri = FString::Printf(TEXT("api/v1/c/ddc/%s/%s/%s"), *Namespace, *DefaultBucket, CacheKey); int64 ResponseCode = 0; uint32 Attempts = 0; // Retry request until we get an accepted response or exhaust allowed number of attempts. while (ResponseCode == 0 && ++Attempts < UE_HTTPDDC_MAX_ATTEMPTS) { if (ShouldAbortForShutdown()) { return EPutStatus::NotCached; } FScopedHttpPoolRequestPtr Request(PutRequestPools[IsInGameThread()].Get()); if (Request.IsValid()) { // Append the content hash to the header HashPayload(Request.Get(), InData); Request->PerformBlockingPut(*Uri, FCompositeBuffer(FSharedBuffer::MakeView(InData.GetData(), InData.Num())), EHttpContentType::Binary); ResponseCode = Request->GetResponseCode(); if (FHttpRequest::IsSuccessResponse(ResponseCode)) { TRACE_COUNTER_ADD(HttpDDC_BytesSent, int64(Request->GetBytesSent())); COOK_STAT(Timer.AddHit(Request->GetBytesSent())); return EPutStatus::Cached; } if (!ShouldRetryOnError(ResponseCode)) { return EPutStatus::NotCached; } ResponseCode = 0; } } #endif // WITH_DATAREQUEST_HELPER return EPutStatus::NotCached; } void FHttpCacheStore::RemoveCachedData(const TCHAR* CacheKey, bool bTransient) { // do not remove transient data as Jupiter does its own verification of the content and cleans itself up if (!IsWritable() || bTransient) return; TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_Remove); FString Uri = FString::Printf(TEXT("api/v1/c/ddc/%s/%s/%s"), *Namespace, *DefaultBucket, CacheKey); int64 ResponseCode = 0; uint32 Attempts = 0; // Retry request until we get an accepted response or exhaust allowed number of attempts. while (ResponseCode == 0 && ++Attempts < UE_HTTPDDC_MAX_ATTEMPTS) { FScopedHttpPoolRequestPtr Request(PutRequestPools[IsInGameThread()].Get()); if (Request.IsValid()) { FHttpRequest::EResult Result = Request->PerformBlockingDelete(*Uri, {}); ResponseCode = Request->GetResponseCode(); if (ResponseCode == 200) { return; } if (!ShouldRetryOnError(ResponseCode)) { return; } ResponseCode = 0; } } } TSharedRef FHttpCacheStore::GatherUsageStats() const { TSharedRef StatsNode = MakeShared(TEXT("Horde Storage"), FString::Printf(TEXT("%s (%s)"), *Domain, *Namespace), /*bIsLocal*/ false); StatsNode->UsageStats.Add(TEXT(""), UsageStats); return StatsNode; } void FHttpCacheStore::Put( const TConstArrayView Requests, IRequestOwner& Owner, FOnCachePutComplete&& OnComplete) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_Put); FRequestBarrier Barrier(Owner); TRefCountedUniqueFunction* CompletionFunction = new TRefCountedUniqueFunction(MoveTemp(OnComplete)); TRefCountPtr> BatchOnCompleteRef(CompletionFunction); for (const FCachePutRequest& Request : Requests) { PutCacheRecordAsync(Owner, Request.Name, Request.Record, Request.Policy, Request.UserData, [COOK_STAT(Timer = UsageStats.TimePut(), ) OnCompletePtr = TRefCountPtr>(CompletionFunction)](FCachePutResponse&& Response, uint64 BytesSent) mutable { TRACE_COUNTER_ADD(HttpDDC_BytesSent, BytesSent); if (Response.Status == EStatus::Ok) { COOK_STAT(if (BytesSent) { Timer.AddHit(BytesSent); }); } OnCompletePtr->GetFunction()(MoveTemp(Response)); }); } } void FHttpCacheStore::Get( const TConstArrayView Requests, IRequestOwner& Owner, FOnCacheGetComplete&& OnComplete) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_Get); FRequestBarrier Barrier(Owner); TRefCountedUniqueFunction* CompletionFunction = new TRefCountedUniqueFunction(MoveTemp(OnComplete)); TRefCountPtr> BatchOnCompleteRef(CompletionFunction); for (const FCacheGetRequest& Request : Requests) { GetCacheRecordAsync(Owner, Request.Name, Request.Key, Request.Policy, Request.UserData, [COOK_STAT(Timer = UsageStats.TimePut(), ) OnCompletePtr = TRefCountPtr>(CompletionFunction)](FCacheGetResponse&& Response, uint64 BytesReceived) mutable { TRACE_COUNTER_ADD(HttpDDC_BytesReceived, BytesReceived); if (Response.Status == EStatus::Ok) { COOK_STAT(Timer.AddHit(BytesReceived);); } OnCompletePtr->GetFunction()(MoveTemp(Response)); }); } } void FHttpCacheStore::PutValue( const TConstArrayView Requests, IRequestOwner& Owner, FOnCachePutValueComplete&& OnComplete) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_PutValue); FRequestBarrier Barrier(Owner); TRefCountedUniqueFunction* CompletionFunction = new TRefCountedUniqueFunction(MoveTemp(OnComplete)); TRefCountPtr> BatchOnCompleteRef(CompletionFunction); for (const FCachePutValueRequest& Request : Requests) { PutCacheValueAsync(Owner, Request.Name, Request.Key, Request.Value, Request.Policy, Request.UserData, [COOK_STAT(Timer = UsageStats.TimePut(),) OnCompletePtr = TRefCountPtr>(CompletionFunction)](FCachePutValueResponse&& Response, uint64 BytesSent) mutable { TRACE_COUNTER_ADD(HttpDDC_BytesSent, BytesSent); if (Response.Status == EStatus::Ok) { COOK_STAT(if (BytesSent) { Timer.AddHit(BytesSent); }); } OnCompletePtr->GetFunction()(MoveTemp(Response)); }); } } void FHttpCacheStore::GetValue( const TConstArrayView Requests, IRequestOwner& Owner, FOnCacheGetValueComplete&& OnComplete) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_GetValue); COOK_STAT(double StartTime = FPlatformTime::Seconds()); COOK_STAT(bool bIsInGameThread = IsInGameThread()); bool bBatchExistsCandidate = true; for (const FCacheGetValueRequest& Request : Requests) { if (!EnumHasAnyFlags(Request.Policy, ECachePolicy::SkipData)) { bBatchExistsCandidate = false; break; } } if (bBatchExistsCandidate) { RefCachedDataProbablyExistsBatchAsync(Owner, Requests, [this, COOK_STAT(StartTime, bIsInGameThread, ) OnComplete = MoveTemp(OnComplete)](FCacheGetValueResponse&& Response) { if (Response.Status != EStatus::Ok) { COOK_STAT(UsageStats.GetStats.Accumulate(FCookStats::CallStats::EHitOrMiss::Miss, FCookStats::CallStats::EStatType::Counter, 1l, bIsInGameThread)); OnComplete(MoveTemp(Response)); } else { UE_LOG(LogDerivedDataCache, Verbose, TEXT("%s: Cache hit for %s from '%s'"), *GetName(), *WriteToString<96>(Response.Key), *Response.Name); COOK_STAT(UsageStats.GetStats.Accumulate(FCookStats::CallStats::EHitOrMiss::Hit, FCookStats::CallStats::EStatType::Counter, 1l, bIsInGameThread)); OnComplete(MoveTemp(Response)); } COOK_STAT(const int64 CyclesUsed = int64((FPlatformTime::Seconds() - StartTime) / FPlatformTime::GetSecondsPerCycle())); COOK_STAT(UsageStats.GetStats.Accumulate(FCookStats::CallStats::EHitOrMiss::Hit, FCookStats::CallStats::EStatType::Cycles, CyclesUsed, bIsInGameThread)); }); } else { FRequestBarrier Barrier(Owner); TRefCountedUniqueFunction* CompletionFunction = new TRefCountedUniqueFunction(MoveTemp(OnComplete)); TRefCountPtr> BatchOnCompleteRef(CompletionFunction); int64 HitBytes = 0; for (const FCacheGetValueRequest& Request : Requests) { GetCacheValueAsync(Owner, Request.Name, Request.Key, Request.Policy, Request.UserData, [this, COOK_STAT(StartTime, bIsInGameThread,) Policy = Request.Policy, OnCompletePtr = TRefCountPtr>(CompletionFunction)] (FCacheGetValueResponse&& Response) { const FOnCacheGetValueComplete& OnComplete = OnCompletePtr->GetFunction(); check(OnComplete); if (Response.Status != EStatus::Ok) { COOK_STAT(UsageStats.GetStats.Accumulate(FCookStats::CallStats::EHitOrMiss::Miss, FCookStats::CallStats::EStatType::Counter, 1l, bIsInGameThread)); OnComplete(MoveTemp(Response)); } else { if (!IsValueDataReady(Response.Value, Policy) && !EnumHasAnyFlags(Policy, ECachePolicy::SkipData)) { // With inline fetching, expect we will always have a value we can use. Even SkipData/Exists can rely on the blob existing if the ref is reported to exist. UE_LOG(LogDerivedDataCache, Log, TEXT("%s: Cache miss due to inlining failure for %s from '%s'"), *GetName(), *WriteToString<96>(Response.Key), *Response.Name); COOK_STAT(UsageStats.GetStats.Accumulate(FCookStats::CallStats::EHitOrMiss::Miss, FCookStats::CallStats::EStatType::Counter, 1l, bIsInGameThread)); OnComplete(MoveTemp(Response)); } else { UE_LOG(LogDerivedDataCache, Verbose, TEXT("%s: Cache hit for %s from '%s'"), *GetName(), *WriteToString<96>(Response.Key), *Response.Name); uint64 ValueSize = Response.Value.GetData().GetCompressedSize(); TRACE_COUNTER_ADD(HttpDDC_BytesReceived, ValueSize); COOK_STAT(UsageStats.GetStats.Accumulate(FCookStats::CallStats::EHitOrMiss::Hit, FCookStats::CallStats::EStatType::Counter, 1l, bIsInGameThread)); OnComplete({ Response.Name, Response.Key, Response.Value, Response.UserData, EStatus::Ok }); COOK_STAT(UsageStats.GetStats.Accumulate(FCookStats::CallStats::EHitOrMiss::Hit, FCookStats::CallStats::EStatType::Bytes, ValueSize, bIsInGameThread)); } } COOK_STAT(const int64 CyclesUsed = int64((FPlatformTime::Seconds() - StartTime) / FPlatformTime::GetSecondsPerCycle())); COOK_STAT(UsageStats.GetStats.Accumulate(FCookStats::CallStats::EHitOrMiss::Hit, FCookStats::CallStats::EStatType::Cycles, CyclesUsed, bIsInGameThread)); }); } } } void FHttpCacheStore::GetChunks( const TConstArrayView Requests, IRequestOwner& Owner, FOnCacheGetChunkComplete&& OnComplete) { TRACE_CPUPROFILER_EVENT_SCOPE(HttpDDC_GetChunks); // TODO: This is inefficient because Jupiter doesn't allow us to get only part of a compressed blob, so we have to // get the whole thing and then decompress only the portion we need. Furthermore, because there is no propagation // between cache stores during chunk requests, the fetched result won't end up in the local store. // These efficiency issues will be addressed by changes to the Hierarchy that translate chunk requests that // are missing in local/fast stores and have to be retrieved from slow stores into record requests instead. That // will make this code path unused/uncommon as Jupiter will most always be a slow store with a local/fast store in front of it. // Regardless, to adhere to the functional contract, this implementation must exist. TArray> SortedRequests(Requests); SortedRequests.StableSort(TChunkLess()); bool bHasValue = false; FValue Value; FValueId ValueId; FCacheKey ValueKey; FCompressedBuffer ValueBuffer; FCompressedBufferReader ValueReader; EStatus ValueStatus = EStatus::Error; FOptionalCacheRecord Record; for (const FCacheGetChunkRequest& Request : SortedRequests) { const bool bExistsOnly = EnumHasAnyFlags(Request.Policy, ECachePolicy::SkipData); COOK_STAT(auto Timer = bExistsOnly ? UsageStats.TimeProbablyExists() : UsageStats.TimeGet()); if (!(bHasValue && ValueKey == Request.Key && ValueId == Request.Id) || ValueReader.HasSource() < !bExistsOnly) { ValueStatus = EStatus::Error; ValueReader.ResetSource(); ValueKey = {}; ValueId.Reset(); Value.Reset(); bHasValue = false; if (Request.Id.IsValid()) { if (!(Record && Record.Get().GetKey() == Request.Key)) { FCacheRecordPolicyBuilder PolicyBuilder(ECachePolicy::None); PolicyBuilder.AddValuePolicy(Request.Id, Request.Policy); Record.Reset(); FRequestOwner BlockingOwner(EPriority::Blocking); GetCacheRecordOnlyAsync(BlockingOwner, Request.Name, Request.Key, PolicyBuilder.Build(), 0, [&Record](FGetCacheRecordOnlyResponse&& Response) { Record = MoveTemp(Response.Record); }); BlockingOwner.Wait(); } if (Record) { const FValueWithId& ValueWithId = Record.Get().GetValue(Request.Id); bHasValue = ValueWithId.IsValid(); Value = ValueWithId; ValueId = Request.Id; ValueKey = Request.Key; if (IsValueDataReady(Value, Request.Policy)) { ValueReader.SetSource(Value.GetData()); } else { auto IdGetter = [](const FValueWithId& Value) { return FString(WriteToString<16>(Value.GetId())); }; FRequestOwner BlockingOwner(EPriority::Blocking); bool bSucceeded = false; FCompressedBuffer NewBuffer; FGetRecordOp::GetDataBatch(*this, BlockingOwner, Request.Name, Request.Key, ::MakeArrayView({ ValueWithId }), IdGetter, [&bSucceeded, &NewBuffer](FGetRecordOp::FGetCachedDataBatchResponse&& Response) { if (Response.Status == EStatus::Ok) { bSucceeded = true; NewBuffer = MoveTemp(Response.DataBuffer); } }); BlockingOwner.Wait(); if (bSucceeded) { ValueBuffer = MoveTemp(NewBuffer); ValueReader.SetSource(ValueBuffer); } else { ValueBuffer.Reset(); ValueReader.ResetSource(); } } } } else { ValueKey = Request.Key; { FRequestOwner BlockingOwner(EPriority::Blocking); bool bSucceeded = false; GetCacheValueAsync(BlockingOwner, Request.Name, Request.Key, Request.Policy, 0, [&bSucceeded, &Value](FCacheGetValueResponse&& Response) { Value = MoveTemp(Response.Value); bSucceeded = Response.Status == EStatus::Ok; }); BlockingOwner.Wait(); bHasValue = bSucceeded; } if (bHasValue) { if (IsValueDataReady(Value, Request.Policy)) { ValueReader.SetSource(Value.GetData()); } else { auto IdGetter = [](const FValue& Value) { return FString(TEXT("Default")); }; FRequestOwner BlockingOwner(EPriority::Blocking); bool bSucceeded = false; FCompressedBuffer NewBuffer; FGetRecordOp::GetDataBatch(*this, BlockingOwner, Request.Name, Request.Key, ::MakeArrayView({ Value }), IdGetter, [&bSucceeded, &NewBuffer](FGetRecordOp::FGetCachedDataBatchResponse&& Response) { if (Response.Status == EStatus::Ok) { bSucceeded = true; NewBuffer = MoveTemp(Response.DataBuffer); } }); BlockingOwner.Wait(); if (bSucceeded) { ValueBuffer = MoveTemp(NewBuffer); ValueReader.SetSource(ValueBuffer); } else { ValueBuffer.Reset(); ValueReader.ResetSource(); } } } else { ValueBuffer.Reset(); ValueReader.ResetSource(); } } } if (bHasValue) { const uint64 RawOffset = FMath::Min(Value.GetRawSize(), Request.RawOffset); const uint64 RawSize = FMath::Min(Value.GetRawSize() - RawOffset, Request.RawSize); UE_LOG(LogDerivedDataCache, Verbose, TEXT("%s: Cache hit for %s from '%s'"), *GetName(), *WriteToString<96>(Request.Key, '/', Request.Id), *Request.Name); COOK_STAT(Timer.AddHit(!bExistsOnly ? RawSize : 0)); FSharedBuffer Buffer; if (!bExistsOnly) { Buffer = ValueReader.Decompress(RawOffset, RawSize); } const EStatus ChunkStatus = bExistsOnly || Buffer.GetSize() == RawSize ? EStatus::Ok : EStatus::Error; OnComplete({Request.Name, Request.Key, Request.Id, Request.RawOffset, RawSize, Value.GetRawHash(), MoveTemp(Buffer), Request.UserData, ChunkStatus}); continue; } OnComplete(Request.MakeResponse(EStatus::Error)); } } } // UE::DerivedData #endif // WITH_HTTP_DDC_BACKEND namespace UE::DerivedData { ILegacyCacheStore* CreateHttpCacheStore( const TCHAR* NodeName, const TCHAR* ServiceUrl, bool bResolveHostCanonicalName, const TCHAR* Namespace, const TCHAR* StructuredNamespace, const TCHAR* OAuthProvider, const TCHAR* OAuthClientId, const TCHAR* OAuthData, const TCHAR* OAuthScope, const FDerivedDataBackendInterface::ESpeedClass* ForceSpeedClass, EBackendLegacyMode LegacyMode, bool bReadOnly) { #if WITH_HTTP_DDC_BACKEND FHttpCacheStore* Backend = new FHttpCacheStore(ServiceUrl, bResolveHostCanonicalName, Namespace, StructuredNamespace, OAuthProvider, OAuthClientId, OAuthData, OAuthScope, LegacyMode, bReadOnly); if (Backend->IsUsable()) { return Backend; } UE_LOG(LogDerivedDataCache, Warning, TEXT("Node %s could not contact the service (%s), will not use it"), NodeName, ServiceUrl); delete Backend; return nullptr; #else UE_LOG(LogDerivedDataCache, Warning, TEXT("HTTP backend is not yet supported in the current build configuration.")); return nullptr; #endif } FDerivedDataBackendInterface* GetAnyHttpCacheStore( FString& OutDomain, FString& OutOAuthProvider, FString& OutOAuthClientId, FString& OutOAuthSecret, FString& OutOAuthScope, FString& OutNamespace, FString& OutStructuredNamespace) { #if WITH_HTTP_DDC_BACKEND if (FHttpCacheStore* HttpBackend = FHttpCacheStore::GetAny()) { OutDomain = HttpBackend->GetDomain(); OutOAuthProvider = HttpBackend->GetOAuthProvider(); OutOAuthClientId = HttpBackend->GetOAuthClientId(); OutOAuthSecret = HttpBackend->GetOAuthSecret(); OutOAuthScope = HttpBackend->GetOAuthScope(); OutNamespace = HttpBackend->GetNamespace(); OutStructuredNamespace = HttpBackend->GetStructuredNamespace(); return HttpBackend; } return nullptr; #else return nullptr; #endif } } // UE::DerivedData