This article presents a comprehensive evaluation of Secure Capsules Layer (SCL) using DataCapsules for inter-enclave communication. Benchmarks on Intel NUCs examine its performance as a key-value store, replication overhead, and the efficiency gains from circular buffers. Results show reduced communication costs compared to SGX SDK and HotCall, while replication introduces scalability limits. Lambda launch latency, dominated by attestation delays, is analyzed, and a fog robotics case study demonstrates SCL’s potential for scaling distributed, enclave-protected applications in real-world use cases.This article presents a comprehensive evaluation of Secure Capsules Layer (SCL) using DataCapsules for inter-enclave communication. Benchmarks on Intel NUCs examine its performance as a key-value store, replication overhead, and the efficiency gains from circular buffers. Results show reduced communication costs compared to SGX SDK and HotCall, while replication introduces scalability limits. Lambda launch latency, dominated by attestation delays, is analyzed, and a fog robotics case study demonstrates SCL’s potential for scaling distributed, enclave-protected applications in real-world use cases.

Is SCL the Key to Faster, Safer Serverless Apps? Here’s What Benchmarks Say

2025/10/03 05:30

Abstract and I. Introduction

II. Background

III. Paranoid Stateful Lambda

IV. SCL Design

V. Optimizations

VI. PSL with SCL

VII. Implementation

VIII. Evaluation

IX. Related Work

X. Conclusion, Acknowledgment, and References

VIII. EVALUATION

SCL leverages DataCaspules as the data representation to support inter-enclave communication. To quantify the benefits and limitations, we ask: (1) How does SCL perform as a KVS(§VIII-B)? (2) How do circular buffer (§VIII-D), and replication (§VIII-C) affect the overhead? (3) How long does it take to securely launch a PSL task? (§VIII-E) (4) How much does SCL pay to run in-enclave distributed applications(§VIII-F)?

\ A. Experiment Setup

\ We evaluate PSL on fifteen Intel NUCs 7PJYH, equiped with Intel(R) Pentium(R) Silver J5005 CPU @ 1.50GHz with 4 physical cores (4 logical threads). The processor has 96K L1 data cache, a 4MiB L2 cache, and 16GB memory. The machine uses Ubuntu 18.04.5 LTS 64bit with Linux 5.4.0- 1048-azure. We run Asylo version 0.6.2. We report the average of experiments that are conducted 10 times. For each NUC, it runs two PSL threads by default.

\ B. End-To-End Benchmark of SCL

\ Benchmark Design: An end-to-end evaluation of SCL starts the worker sends the first acknowledgement to the user, and ends when the client receives its last request’s response from the workers. We evaluate the performance using a workload generated by YCSB workload generators. Due to the difference between get and put protocols, we focus on the read-only and write-only workloads. All workloads comply zipfian distribution, where keys are accessed at non-uniform frequency. For each get, we evaluate the performance of getting from the local memtable of the lambda(get(cached)), and of getting the data from CapsuleDB(get(uncached)). Each get request is synchronous that the next request is sent only if it gets the value of the previous get request.

\ Overall Performance: Figure 9 shows the throughput of the end-to-end YCSB benchmark. The aggregated throughput of put. The get(CapsuleDB) throughput is flattened as we increate the number of the lambdas, because we run one single CapsuleDB instance that handle all the queries, which is bottlenecked as the number of lambdas that issue get(CapsuleDB) increases.

\ Fig. 9: The aggregated throughput of PSL on put-only and get-only workload of YCSB benchmark. The gets are differentiated by whether it is cached in the memtable(left) or the lambda needs to query CapsuleDB(right).

\ Fig. 10: A line graph that shows end-to-end write-only benchmarks for SCL with vs. without replication. Throughput numbers are in log scale.

\ C. Replication-enabled End-To-End Benchmark

\ Benchmark Design: Replication-enabled end-to-end evaluation measures the performance of the SCL layer with durability. In particular, it includes the overhead of workers sending each write to the DataCapsule replicas, a quorum of DataCapsule replicas receive data and persist it on disk, and then acking the worker. We evaluate the performance using a workload generated by YCSB workload generators. Since replication involves only write operations, we evaluate a writeonly workload. The workload involves a zipfian distribution, with keys accessed at a non-uniform frequency.

\ Overall Performance: Figure 10 illustrates the performance of the DataCapsule backend. It shows that SCL with replication has reached a bottleneck after 9 workers while SCL without durability continues to scale. The performance drop and bottleneck are due to several reasons: 1) disk operations are inherently slow; 2) the burden on replication system’s leader is high for collecting acks from DataCapsule replicas and sending the aggregated ack back to worker. We aim to improve SCL with replication by mitigating the workload on the replication leader.

\ D. Circular Buffer Microbenchmark

\ Benchmark Design: The circular buffer provides efficient application-enclave communication. We compare the perfor-

\ TABLE I: Circular Buffer Microbenchmark We evaluate the number of clock cycles required for communications between the enclave and application in both directions.

\ Fig. 11: Latency breakdown of the Paranoid Stateful Lambda launching process. The bold line represents the critical path of the lambda launching process. The total launching time to run code in authenticated worker is less than 0.61 second.

\ mance of the circular buffer with the SGX SDK baseline and the state-of-the-art HotCall. We evaluate them based on the number of clock cycles required for communications in both the application to enclave direction and vice versa.

\ Overall Performance: As shown in Table I, baseline SGX SDK incurs a significant overhead of over 20,000 clock cycles from application to enclave, and over 8,600 clock cycles from enclave to application. For both directions, HotCall is able to reduce the overhead to under a thousand clock cycles. Our circular buffer reduces overheads even further. Our solution only requires 461.1 and 525.54 clock cycles from application to enclave and vice versa. Compared to state-of-the-art HotCall, our solution provides 103% and 44% improvements, respectively.

\ E. Lambda Launch Time

\ Benchmark Design: We evaluate the launching process of PSL by running Workers and FaaS leader in SGXv2 hardware mode, which the worker lambda, FaaS leader and user on different physical Intel NUCs machines. For each NUC, it runs Asylo AGE in hardware mode with PCE signed by Intel that helps enclave generates attestation assertions. We assume the machines already have the pulled the prebuilt lambda runtime binaries and execute the runtime. The cold-start bootstrapping process lasts 42 seconds on average in our experiment setting.

\ Lambda Launch Breakdown: Figure 11 show the latency breakdown of the launching process. It takes 0.30s for the user to reach out to the scheduler, and for the scheduler to find and forward the encrypted task to the potential workers. Then the workers load associated runtime and data to the enclave, which takes 0.16s. We parallelize the worker loading time with the attestation. that the user remotely attests the FaaS Leader to verify that the FaaS leader is running authenticated code in SGX enclave. After the worker’s enclave file is loaded, it takes 0.103s on average for the FaaS leader to remotely attest the

\ TABLE II: Motion Planning Benchmarks

\ worker enclave. We note that this attestation latency is mostly constituted by the network delay of grpc request and the local attestation assertion generation time of the worker’s AGE, so it does not incur scalability issue with the FaaS leader when multiple workers are launched at the same time.

\ F. Case Study: Fog Robotics Motion Planner

\ We experiment with a sampling-based motion planner that is parallelized to run on multiple concurrent serverless processes, MPLambda [23], and modifying it to use SCL. Most of the porting effort done was to integrate MPLambda’s build system into Asylo. The modification is about 100 LoC. Many system calls that MPLambda uses are proxied by Asylo.

\ Using MPLambda with SCL, we compute a motion plan running a fetch scenario in which a Fetch mobile manipulator robot [17] declutters a desk. We measure the median wallclock time to find the first solution by the planners. We also measure the median average path cost per time of the lowest cost path the planners return. This captures how efficiently the planners can compute the best path. Because the planner uses random sampling, we run the same computation multiple times with different seeds. As with previous experiments, we run this test on Intel NUCs 7PJYH, equipped with Intel(R) Pentium(R) Silver J5005 CPU @ 1.50GHz with 4 physical cores (4 logical threads). We set a timeout of 600 seconds for the planners to compute a path.

\ We run up to 8 planners, running on separate Intel NUCs using SCL and comparing this to running MPLamda without SCL. We observe an increase in performance as we scale out the number of planners. Each planner runs computationally heavy workloads and PSL introduces several threads (i.e. crypto actors, zmq clients, OCALL/ECALL handlers) that take away CPU time from the planner thread. Furthermore, MPLambda planners use the Rapidly-exploring random tree (RRT*) [24] algorithm, to search for paths by randomly generating samples from a search space, checking whether the sample is feasible to explore, and adding the sample to a constructed tree data structure. The tree data structure may grow large and take up a significant amount of memory. Memory in SGX is a limited resource and increased memory pressure leads to more misses in the EPC and requiring paging in and out of enclaves frequently. There is work on limiting the memory usage of RRT* by bounding the memory for the tree data structure, which we can adopt in future work. [2].

\

:::info Authors:

(1) Kaiyuan Chen, University of California, Berkeley (kych@berkeley.edu);

(2) Alexander Thomas, University of California, Berkeley (alexthomas@berkeley.edu);

(3) Hanming Lu, University of California, Berkeley (hanming lu@berkeley.edu);

(4) William Mullen, University of California, Berkeley (wmullen@berkeley.edu);

(5) Jeff Ichnowski, University of California, Berkeley (jeffi@berkeley.edu);

(6) Rahul Arya, University of California, Berkeley (rahularya@berkeley.edu);

(7) Nivedha Krishnakumar, University of California, Berkeley (nivedha@berkeley.edu);

(8) Ryan Teoh, University of California, Berkeley (ryanteoh@berkeley.edu);

(9) Willis Wang, University of California, Berkeley (williswang@berkeley.edu);

(10) Anthony Joseph, University of California, Berkeley (adj@berkeley.edu);

(11) John Kubiatowicz, University of California, Berkeley (kubitron@berkeley.edu).

:::


:::info This paper is available on arxiv under CC BY 4.0 DEED license.

:::

\

Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact service@support.mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.
Share Insights

You May Also Like

Franklin Templeton CEO Dismisses 50bps Rate Cut Ahead FOMC

Franklin Templeton CEO Dismisses 50bps Rate Cut Ahead FOMC

The post Franklin Templeton CEO Dismisses 50bps Rate Cut Ahead FOMC appeared on BitcoinEthereumNews.com. Franklin Templeton CEO Jenny Johnson has weighed in on whether the Federal Reserve should make a 25 basis points (bps) Fed rate cut or 50 bps cut. This comes ahead of the Fed decision today at today’s FOMC meeting, with the market pricing in a 25 bps cut. Bitcoin and the broader crypto market are currently trading flat ahead of the rate cut decision. Franklin Templeton CEO Weighs In On Potential FOMC Decision In a CNBC interview, Jenny Johnson said that she expects the Fed to make a 25 bps cut today instead of a 50 bps cut. She acknowledged the jobs data, which suggested that the labor market is weakening. However, she noted that this data is backward-looking, indicating that it doesn’t show the current state of the economy. She alluded to the wage growth, which she remarked is an indication of a robust labor market. She added that retail sales are up and that consumers are still spending, despite inflation being sticky at 3%, which makes a case for why the FOMC should opt against a 50-basis-point Fed rate cut. In line with this, the Franklin Templeton CEO said that she would go with a 25 bps rate cut if she were Jerome Powell. She remarked that the Fed still has the October and December FOMC meetings to make further cuts if the incoming data warrants it. Johnson also asserted that the data show a robust economy. However, she noted that there can’t be an argument for no Fed rate cut since Powell already signaled at Jackson Hole that they were likely to lower interest rates at this meeting due to concerns over a weakening labor market. Notably, her comment comes as experts argue for both sides on why the Fed should make a 25 bps cut or…
Share
2025/09/18 00:36
Ethereum’s ERC-8004 Brings AI-Driven Economic Potential

Ethereum’s ERC-8004 Brings AI-Driven Economic Potential

The post Ethereum’s ERC-8004 Brings AI-Driven Economic Potential appeared on BitcoinEthereumNews.com. Key Points: ERC-8004 launch by Cobo enables AI as economic entities in crypto. No immediate market impact noted yet. Potential for significant future Ethereum ecosystem evolution. Cobo’s co-founder Fish the Godfish introduced a groundbreaking crypto stack—x402, AP2, and ERC-8004—on September 17th, enabling AI agents to transact as economic entities officially. This technical advancement fosters new machine involvement in economic activities within Ethereum, anticipated to alter future DeFi landscapes, despite no current financial or market impact observed. ERC-8004 and AI: Transforming Ethereum Transactions Cobo’s ERC-8004 aims to transform the cryptocurrency landscape by allowing AI agents to engage in economic activities, introducing a stack that interlinks x402 and AP2 for seamless transactions. Fish the Godfish, the primary architect of this initiative, has highlighted the potential for AI to evolve into true economic agents, changing how transactions are approached in blockchain ecosystems. The introduction of this stack is a technological milestone, though no immediate financial impact has surfaced. The stack positions Ethereum as a hub for machine-led commerce, foreshadowing future changes in decentralized finance and smart contract applications. When AI learns to spend: From x402 to AP2, and then to ERC-8004, explore how to make the Agent a true economic entity. — Fish the Godfish, Co-founder and CEO of Cobo Reactions to the announcement have been cautiously optimistic, with many in the community anticipating advancements, although industry influencers have yet to comment. This caution suggests that while the technical potential is acknowledged, its market and practical impacts remain speculative. Ethereum’s Evolution: AI Agents and Market Dynamics Did you know? ERC-8004, hailed as a significant advancement, has historical parallels with early smart contract technologies that first enabled programmable transactions on blockchains. Ethereum (ETH) is valued at $3,957.24 with a market cap of 477,631,941,155. Its 24-hour trading volume is $15.36 billion, showing a -55.14% change,…
Share
2025/10/26 07:35
XRP (XRP) Faces Potential Downturn as Death Cross Pattern Re-emerges

XRP (XRP) Faces Potential Downturn as Death Cross Pattern Re-emerges

The post XRP (XRP) Faces Potential Downturn as Death Cross Pattern Re-emerges appeared on BitcoinEthereumNews.com. Ted Hisokawa Oct 24, 2025 16:07 XRP is on the brink of forming a ‘death cross’ pattern, reminiscent of its 65% crash in 2021. Experts warn of potential risks including falling burn rate and insider selling. The price of XRP, the cryptocurrency developed by Ripple, is currently navigating a challenging phase, marked by a significant decline from its peak earlier this year. According to CoinMarketCap, XRP has dropped by 34% from its highest point, situating it firmly within a bearish market. Death Cross Pattern and Historical Context A looming ‘death cross’ pattern on the daily chart is raising alarms among analysts. This technical chart pattern, which occurs when a short-term moving average crosses below a long-term moving average, has historically signaled a potential downturn. The last instance of this pattern for XRP was in 2021, leading to a dramatic 65% price drop. Current Market Conditions As of October 23, XRP was trading at $2.4137, a price level that reflects recent volatility and market consolidation. This price action is consistent with broader trends observed across the altcoin market, where significant price swings have been common since early October. Despite these challenges, XRP remains a key player in the cryptocurrency space, backed by robust fundamentals. Additional Risks for XRP Beyond the technical patterns, XRP faces other risks that could impact its price. Notably, the burn rate for the token is declining, which could affect its perceived scarcity and value. Furthermore, insider selling has been flagged as a potential concern, possibly contributing to downward pressure on the price. Market Developments and Future Outlook In contrast to the current bearish sentiment, Ripple’s ecosystem continues to expand. The recent launch of the REX-Oprey XRP ETF has been a significant milestone, quickly surpassing $100 million in assets. This…
Share
2025/10/26 07:24