https://fred.stlouisfed.org/legal/#copyright-public-domainhttps://fred.stlouisfed.org/legal/#copyright-public-domain
View data of PCE, an index that measures monthly changes in the price of consumer goods and services as a means of analyzing inflation.
https://fred.stlouisfed.org/legal/#copyright-public-domainhttps://fred.stlouisfed.org/legal/#copyright-public-domain
Graph and download economic data for Personal Consumption Expenditures Excluding Food and Energy (Chain-Type Price Index) (PCEPILFE) from Jan 1959 to May 2025 about chained, core, energy, headline figure, PCE, consumption expenditures, consumption, personal, inflation, price index, indexes, price, and USA.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
PCE Price Index Annual Change in the United States increased to 2.30 percent in May from 2.20 percent in April of 2025. This dataset includes a chart with historical data for the United States PCE Price Index Annual Change.
https://fred.stlouisfed.org/legal/#copyright-public-domainhttps://fred.stlouisfed.org/legal/#copyright-public-domain
Graph and download economic data for Personal Consumption Expenditures: Chain-type Price Index (PCEPI) from Jan 1959 to May 2025 about chained, headline figure, PCE, consumption expenditures, consumption, personal, inflation, price index, indexes, price, and USA.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
PCE Prices QoQ in the United States increased to 3.70 percent in the first quarter of 2025 from 2.40 percent in the fourth quarter of 2024. This dataset includes a chart with historical data for the United States PCE Prices QoQ.
Personal consumption expenditures (PCE) is the value of the goods and services purchased by, or on the behalf of, Iowa residents. Per capita PCE is calculated by dividing the PCE by the Census Bureau’s annual midyear (July 1) population estimates.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Core PCE Price Index Annual Change in the United States increased to 2.70 percent in May from 2.60 percent in April of 2025. This dataset includes a chart with historical data for the United States Core Pce Price Index Annual Change.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Nowcast for Real Personal Consumption Expenditures uses a nowcasting model to synthesize the bridge equation approach relating GDP subcomponents to monthly source data with factor model and Bayesian vector autoregression approaches.
The Federal Reserve Bank of Atlanta’s GDPNow release complements the quarterly GDP release from the Bureau of Economic Analysis (BEA). The Atlanta Fed recalculates and updates their GDPNow forecasts (called “nowcasts”) throughout the quarter as new data are released, up until the BEA releases its “advance estimate” of GDP for that quarter. The St. Louis Fed constructs a quarterly time series for this dataset, in which both historical and current observations values are combined. In general, the most-current observation is revised multiple times throughout the quarter. The final forecasted value (before the BEA’s release of the advance estimate of GDP) is the static, historical value for that quarter.
For futher information visit the source at https://www.frbatlanta.org/cqer/research/gdpnow.aspx?panel=1.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset is related to the 100D function benchmark case. A detailed description of the benchmark case can be found on the public online community website UQWorld: https://uqworld.org/t/benchmark-case-100d-function/.
The experimental designs include datasets with 400, 800, 1200, 1600, and 2000 samples, each generated using optimized maximin distance Latin Hypercube Sampling (LHS) with 1000 iterations. Each dataset is replicated 20 times. The validation set contains 100,000 samples generated by Monte Carlo simulation. Each dataset contains input samples and the corresponding computational model responses.
The dataset file includes two variables:
Both variables are Matlab structures with fields X, Y, and nSamples. Variable ExpDesigns is a non-scalar structure sized according to the number of experimental design groups. Each field of X for the i-th element of the struct array contains replicated datasets, forming a matrix of size [number of samples] x [dimensionality] x [number of replications]. Similarly, each field of Y for the i-th element contains replicated computational model responses that correspond to the experimental design of the same replication, sized [number of samples] x [number of model outputs] x [number of replications]. The same structure logic applies to the ValidationSet variable, except it contains only one dataset per benchmark case.
The structure can be summarized as follows:
The selection of competitors was based on our experience with meta-modeling and includes various metamodel types: Polynomial Chaos Expansions (PCE), Polynomial Chaos Kriging (PCK), and Kriging. Given that each metamodel has many hyperparameters, we chose the most general settings to address different benchmark case difficulties, including dimensionality, nonlinearity, and non-monotonicity.
For Polynomial Chaos Expansions (PCE), we used a polynomial degree and q-norm adaptivity approach. This approach adaptively increases the maximum polynomial degree and truncation q-norm until the estimated leave-one-out error starts increasing. Maximum polynomial interaction terms were limited to 2 due to the memory requirements for large model dimensionality and large experimental designs. We tested three different solvers to calculate the PCE coefficients: Least Angle Regression (LARS), Orthogonal Matching Pursuit (OMP), and Subspace Pursuit (SP).
Polynomial Chaos Kriging (PCK) employs a sequential combination strategy of PCE and Kriging. PCE uses degree adaptivity with a fixed q-norm. The maximum number of interactions is again set to 2 with the LARS solver. Ordinary Kriging is applied using the Matérn-5/2 correlation family, ellipsoidal, and anisotropic correlation function. We used a hybrid genetic algorithm to optimize the hyperparameters.
We benchmarked both linear and ordinary Kriging, including Matérn-5/2 and Gaussian correlation families and separable and ellipsoidal correlation, resulting in eight different Kriging competitors. The hyperparameters were calculated using a hybrid covariance matrix adaptation-evolution strategy optimization.
For further details on the settings, please refer to the competitors.m file and UQLab user manuals:
The results file contains one variable: Metrics. It is a Matlab structure with fields corresponding to each competitor (currently 12). Each competitor field contains data of type non-scalar struct array. The performance metrics included are RelMSE, RelRMSE, RelMAE, MAPE, Q2, and RelCVErr. Each field of Metrics.(CompetitorName) for the i-th element of the struct array contains metrics corresponding to the replicated dataset and the competitor, structured as follows:
The description of the performance measures (metrics) can be found here: https://uqworld.org/t/metamodel-performance-measures/.
We provide files in three languages (MATLAB, Python, and Julia) to showcase how to work with datasets, results, and their visualization. The files are called working_with_datafiles.* (the extension depends on the selected language).
This project was supported by the Open Research Data Program of the ETH Board under Grant number EPFL SCR0902285. The calculations were run on the Euler cluster of ETH Zürich using the MATLAB-based UQLab software developed at the Chair of Risk, Safety and Uncertainty Quantification of ETH Zürich.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset is related to the Two-dimensional heat diffusion model benchmark case. A detailed description of the benchmark case can be found on the public online community website UQWorld: https://uqworld.org/t/benchmark-case-two-dimensional-heat-diffusion-model/.
The experimental designs include datasets with 400, 800, 1200, 1600, and 2000 samples, each generated using optimized maximin distance Latin Hypercube Sampling (LHS) with 1000 iterations. Each dataset is replicated 20 times. The validation set contains 100,000 samples generated by Monte Carlo simulation. Each dataset contains input samples and the corresponding computational model responses.
The dataset file includes two variables:
Both variables are Matlab structures with fields X, Y, and nSamples. Variable ExpDesigns is a non-scalar structure sized according to the number of experimental design groups. Each field of X for the i-th element of the struct array contains replicated datasets, forming a matrix of size [number of samples] x [dimensionality] x [number of replications]. Similarly, each field of Y for the i-th element contains replicated computational model responses that correspond to the experimental design of the same replication, sized [number of samples] x [number of model outputs] x [number of replications]. The same structure logic applies to the ValidationSet variable, except it contains only one dataset per benchmark case.
The structure can be summarized as follows:
The selection of competitors was based on our experience with meta-modeling and includes various metamodel types: Polynomial Chaos Expansions (PCE), Polynomial Chaos Kriging (PCK), and Kriging. Given that each metamodel has many hyperparameters, we chose the most general settings to address different benchmark case difficulties, including dimensionality, nonlinearity, and non-monotonicity.
For Polynomial Chaos Expansions (PCE), we used a polynomial degree and q-norm adaptivity approach. This approach adaptively increases the maximum polynomial degree and truncation q-norm until the estimated leave-one-out error starts increasing. Maximum polynomial interaction terms were limited to 2 due to the memory requirements for large model dimensionality and large experimental designs. We tested three different solvers to calculate the PCE coefficients: Least Angle Regression (LARS), Orthogonal Matching Pursuit (OMP), and Subspace Pursuit (SP).
Polynomial Chaos Kriging (PCK) employs a sequential combination strategy of PCE and Kriging. PCE uses degree adaptivity with a fixed q-norm. The maximum number of interactions is again set to 2 with the LARS solver. Ordinary Kriging is applied using the Matérn-5/2 correlation family, ellipsoidal, and anisotropic correlation function. We used a hybrid genetic algorithm to optimize the hyperparameters.
We benchmarked both linear and ordinary Kriging, including Matérn-5/2 and Gaussian correlation families and separable and ellipsoidal correlation, resulting in eight different Kriging competitors. The hyperparameters were calculated using a hybrid covariance matrix adaptation-evolution strategy optimization.
For further details on the settings, please refer to the competitors.m file and UQLab user manuals:
The results file contains one variable: Metrics. It is a Matlab structure with fields corresponding to each competitor (currently 12). Each competitor field contains data of type non-scalar struct array. The performance metrics included are RelMSE, RelRMSE, RelMAE, MAPE, Q2, and RelCVErr. Each field of Metrics.(CompetitorName) for the i-th element of the struct array contains metrics corresponding to the replicated dataset and the competitor, structured as follows:
The description of the performance measures (metrics) can be found here: https://uqworld.org/t/metamodel-performance-measures/.
We provide files in three languages (MATLAB, Python, and Julia) to showcase how to work with datasets, results, and their visualization. The files are called working_with_datafiles.* (the extension depends on the selected language).
This project was supported by the Open Research Data Program of the ETH Board under Grant number EPFL SCR0902285. The calculations were run on the Euler cluster of ETH Zürich using the MATLAB-based UQLab software developed at the Chair of Risk, Safety and Uncertainty Quantification of ETH Zürich.
https://fred.stlouisfed.org/legal/#copyright-public-domainhttps://fred.stlouisfed.org/legal/#copyright-public-domain
Graph and download economic data for Personal Saving Rate (PSAVERT) from Jan 1959 to May 2025 about savings, personal, rate, and USA.
Not seeing a result you expected?
Learn how you can add new datasets to our index.
https://fred.stlouisfed.org/legal/#copyright-public-domainhttps://fred.stlouisfed.org/legal/#copyright-public-domain
View data of PCE, an index that measures monthly changes in the price of consumer goods and services as a means of analyzing inflation.