Facebook
TwitterCC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
analyze the current population survey (cps) annual social and economic supplement (asec) with r the annual march cps-asec has been supplying the statistics for the census bureau's report on income, poverty, and health insurance coverage since 1948. wow. the us census bureau and the bureau of labor statistics ( bls) tag-team on this one. until the american community survey (acs) hit the scene in the early aughts (2000s), the current population survey had the largest sample size of all the annual general demographic data sets outside of the decennial census - about two hundred thousand respondents. this provides enough sample to conduct state- and a few large metro area-level analyses. your sample size will vanish if you start investigating subgroups b y state - consider pooling multiple years. county-level is a no-no. despite the american community survey's larger size, the cps-asec contains many more variables related to employment, sources of income, and insurance - and can be trended back to harry truman's presidency. aside from questions specifically asked about an annual experience (like income), many of the questions in this march data set should be t reated as point-in-time statistics. cps-asec generalizes to the united states non-institutional, non-active duty military population. the national bureau of economic research (nber) provides sas, spss, and stata importation scripts to create a rectangular file (rectangular data means only person-level records; household- and family-level information gets attached to each person). to import these files into r, the parse.SAScii function uses nber's sas code to determine how to import the fixed-width file, then RSQLite to put everything into a schnazzy database. you can try reading through the nber march 2012 sas importation code yourself, but it's a bit of a proc freak show. this new github repository contains three scripts: 2005-2012 asec - download all microdata.R down load the fixed-width file containing household, family, and person records import by separating this file into three tables, then merge 'em together at the person-level download the fixed-width file containing the person-level replicate weights merge the rectangular person-level file with the replicate weights, then store it in a sql database create a new variable - one - in the data table 2012 asec - analysis examples.R connect to the sql database created by the 'download all microdata' progr am create the complex sample survey object, using the replicate weights perform a boatload of analysis examples replicate census estimates - 2011.R connect to the sql database created by the 'download all microdata' program create the complex sample survey object, using the replicate weights match the sas output shown in the png file below 2011 asec replicate weight sas output.png statistic and standard error generated from the replicate-weighted example sas script contained in this census-provided person replicate weights usage instructions document. click here to view these three scripts for more detail about the current population survey - annual social and economic supplement (cps-asec), visit: the census bureau's current population survey page the bureau of labor statistics' current population survey page the current population survey's wikipedia article notes: interviews are conducted in march about experiences during the previous year. the file labeled 2012 includes information (income, work experience, health insurance) pertaining to 2011. when you use the current populat ion survey to talk about america, subract a year from the data file name. as of the 2010 file (the interview focusing on america during 2009), the cps-asec contains exciting new medical out-of-pocket spending variables most useful for supplemental (medical spending-adjusted) poverty research. confidential to sas, spss, stata, sudaan users: why are you still rubbing two sticks together after we've invented the butane lighter? time to transition to r. :D
Facebook
Twitter
According to our latest research, the global SAS HBA (Serial Attached SCSI Host Bus Adapter) market size reached USD 1.47 billion in 2024, and it is poised to grow at a CAGR of 7.1% during the forecast period, reaching an estimated USD 2.74 billion by 2033. This robust growth is driven by increasing demand for high-speed and reliable data transfer solutions across data centers, enterprise storage, and server environments. The proliferation of big data analytics, cloud computing, and the expansion of enterprise IT infrastructure are among the primary factors fueling market expansion, as organizations worldwide seek efficient and scalable storage connectivity solutions.
One of the most significant growth factors for the SAS HBA market is the exponential rise in data generation and storage requirements across various industries. With digital transformation initiatives accelerating globally, organizations are investing heavily in advanced storage systems to manage and process vast volumes of data efficiently. SAS HBAs play a crucial role in enabling high-speed, low-latency connections between servers and storage devices, ensuring seamless data flow and robust performance. The growing adoption of cloud-based services, virtualization, and high-performance computing (HPC) further amplifies the need for scalable and reliable storage connectivity, driving the demand for SAS HBA solutions in both enterprise and hyperscale data center environments.
Another critical driver propelling the SAS HBA market is the ongoing evolution of storage technologies and the increasing complexity of enterprise IT infrastructure. As businesses transition from traditional storage architectures to more sophisticated, hybrid, and software-defined storage environments, the need for versatile and high-capacity connectivity solutions has become paramount. SAS HBAs offer backward compatibility, enhanced error correction, and superior scalability compared to legacy solutions, making them an ideal choice for organizations seeking to future-proof their storage investments. The integration of advanced features such as multi-path I/O, improved power management, and support for higher data transfer rates positions SAS HBAs as essential components in modern IT ecosystems.
Furthermore, the surge in demand for mission-critical applications and real-time data processing across sectors such as BFSI, healthcare, manufacturing, and government is accelerating the adoption of SAS HBA solutions. These applications require uninterrupted access to large datasets and depend on the high reliability and performance provided by SAS HBA technology. The increasing prevalence of AI, machine learning, and IoT-driven workloads is also contributing to the marketÂ’s momentum, as these technologies necessitate robust storage connectivity to handle intensive data processing requirements. As a result, vendors are continuously innovating and expanding their product portfolios to cater to the evolving needs of diverse end-users.
In addition to SAS HBAs, Fibre Channel HBA technology is gaining traction as an alternative storage connectivity solution, particularly in environments where high-speed data transfer and low latency are critical. Fibre Channel HBAs are known for their ability to provide dedicated bandwidth and enhanced reliability, making them a preferred choice for mission-critical applications in sectors such as finance, healthcare, and telecommunications. As organizations continue to seek robust and scalable storage solutions, the integration of Fibre Channel HBAs into existing IT infrastructures offers a pathway to achieving optimal performance and efficiency. The growing adoption of this technology underscores the importance of versatile connectivity options in modern data center environments.
From a regional perspective, North America continues to dominate the global SAS HBA market, accounting for the largest revenue share in 2024, followed by Europe and the Asia Pacific. The strong presence of leading technology companies, early adoption of advanced storage solutions, and significant investments in data center infrastructure are key factors supporting North AmericaÂ’s leadership position. Meanwhile, the Asia Pacific region is witnessing the fastest growth, driven by rapid digitalization, expanding enterprise IT infrastructure, and increasing investment
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Total-Long-Term-Liabilities Time Series for Believe SAS. Believe S.A. provides digital music services for independent labels and local artists in France, Germany, rest of Europe, the Americas, Asia, Oceania, and Pacific. It operates through two segments, Premium Solutions and Automated Solutions. The company engages in the sale, promotion, and delivery of digital content provided by artists and labels by developing their catalog on digital platforms and social media; administration of copyrights; provision of synchronization services comprising the use of recorded music in advertising, films and series, video games and television; and organization of musical events. It also offers TuneCore digital platform for artists to distribute their audio content in an automated manner to streaming and social media platforms. Believe S.A. was incorporated in 2005 and is headquartered in Paris, France.
Facebook
TwitterThis database is a collection of maps created from the 28 SAS-2 observation files. The original observation files can be accessed within BROWSE by changing to the SAS2RAW database. For each of the SAS-2 observation files, the analysis package FADMAP was run and the resulting maps, plus GIF images created from these maps, were collected into this database. Each map is a 60 x 60 pixel FITS format image with 1 degree pixels. The user may reconstruct any of these maps within the captive account by running FADMAP from the command line after extracting a file from within the SAS2RAW database. The parameters used for selecting data for these product map files are embedded keywords in the FITS maps themselves. These parameters are set in FADMAP, and for the maps in this database are set as 'wide open' as possible. That is, except for selecting on each of 3 energy ranges, all other FADMAP parameters were set using broad criteria. To find more information about how to run FADMAP on the raw event's file, the user can access help files within the SAS2RAW database or can use the 'fhelp' facility from the command line to gain information about FADMAP. This is a service provided by NASA HEASARC .
Facebook
TwitterIn this paper, we investigate the use of Bayesian networks to construct large-scale diagnostic systems. In particular, we consider the development of large-scale Bayesian networks by composition. This compositional approach reflects how (often redundant) subsystems are architected to form systems such as electrical power systems. We develop high-level specifications, Bayesian networks, clique trees, and arithmetic circuits representing 24 different electrical power systems. The largest among these 24 Bayesian networks contains over 1,000 random variables. Another BN represents the real-world electrical power system ADAPT, which is representative of electrical power systems deployed in aerospace vehicles. In addition to demonstrating the scalability of the compositional approach, we briefly report on experimental results from the diagnostic competition DXC, where the ProADAPT team, using techniques discussed here, obtained the highest scores in both Tier 1 (among 9 international competitors) and Tier 2 (among 6 international competitors) of the industrial track. While we consider diagnosis of power systems specically, we believe this work is relevant to other system health management problems, in particular in dependable systems such as aircraft and spacecraft. Reference: O. J. Mengshoel, S. Poll, and T. Kurtoglu. "Developing Large-Scale Bayesian Networks by Composition: Fault Diagnosis of Electrical Power Systems in Aircraft and Spacecraft." Proc. of the IJCAI-09 Workshop on Self-* and Autonomous Systems (SAS): Reasoning and Integration Challenges, 2009 BibTex Reference: @inproceedings{mengshoel09developing, title = {Developing Large-Scale {Bayesian} Networks by Composition: Fault Diagnosis of Electrical Power Systems in Aircraft and Spacecraft}, author = {Mengshoel, O. J. and Poll, S. and Kurtoglu, T.}, booktitle = {Proc. of the IJCAI-09 Workshop on Self-$\star$ and Autonomous Systems (SAS): Reasoning and Integration Challenges}, year={2009} }
Facebook
TwitterCC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
NIS 2002-2011 Within Year Merge
Facebook
Twitterhttps://www.marketresearchforecast.com/privacy-policyhttps://www.marketresearchforecast.com/privacy-policy
The global Data Mining Software market is experiencing robust growth, driven by the increasing need for businesses to extract valuable insights from massive datasets. The market, estimated at $15 billion in 2025, is projected to witness a Compound Annual Growth Rate (CAGR) of 12% from 2025 to 2033, reaching an estimated $45 billion by 2033. This expansion is fueled by several key factors. The burgeoning adoption of cloud-based solutions offers scalability and cost-effectiveness, attracting both large enterprises and SMEs. Furthermore, advancements in machine learning and artificial intelligence algorithms are enhancing the accuracy and efficiency of data mining processes, leading to better decision-making across various sectors like finance, healthcare, and marketing. The rise of big data analytics and the increasing availability of affordable, high-powered computing resources are also significant contributors to market growth. However, the market faces certain challenges. Data security and privacy concerns remain paramount, especially with the increasing volume of sensitive information being processed. The complexity of data mining software and the need for skilled professionals to operate and interpret the results present a barrier to entry for some businesses. The high initial investment cost associated with implementing sophisticated data mining solutions can also deter smaller organizations. Nevertheless, the ongoing technological advancements and the growing recognition of the strategic value of data-driven decision-making are expected to overcome these restraints and propel the market toward continued expansion. The market segmentation reveals a strong preference for cloud-based solutions, reflecting the industry's trend toward flexible and scalable IT infrastructure. Large enterprises currently dominate the market share, but SMEs are rapidly adopting data mining software, indicating promising future growth in this segment. Geographic analysis shows that North America and Europe are currently leading the market, but the Asia-Pacific region is poised for significant growth due to increasing digitalization and economic expansion in countries like China and India.
Facebook
Twitteranalyze the health and retirement study (hrs) with r the hrs is the one and only longitudinal survey of american seniors. with a panel starting its third decade, the current pool of respondents includes older folks who have been interviewed every two years as far back as 1992. unlike cross-sectional or shorter panel surveys, respondents keep responding until, well, death d o us part. paid for by the national institute on aging and administered by the university of michigan's institute for social research, if you apply for an interviewer job with them, i hope you like werther's original. figuring out how to analyze this data set might trigger your fight-or-flight synapses if you just start clicking arou nd on michigan's website. instead, read pages numbered 10-17 (pdf pages 12-19) of this introduction pdf and don't touch the data until you understand figure a-3 on that last page. if you start enjoying yourself, here's the whole book. after that, it's time to register for access to the (free) data. keep your username and password handy, you'll need it for the top of the download automation r script. next, look at this data flowchart to get an idea of why the data download page is such a righteous jungle. but wait, good news: umich recently farmed out its data management to the rand corporation, who promptly constructed a giant consolidated file with one record per respondent across the whole panel. oh so beautiful. the rand hrs files make much of the older data and syntax examples obsolete, so when you come across stuff like instructions on how to merge years, you can happily ignore them - rand has done it for you. the health and retirement study only includes noninstitutionalized adults when new respondents get added to the panel (as they were in 1992, 1993, 1998, 2004, and 2010) but once they're in, they're in - respondents have a weight of zero for interview waves when they were nursing home residents; but they're still responding and will continue to contribute to your statistics so long as you're generalizing about a population from a previous wave (for example: it's possible to compute "among all americans who were 50+ years old in 1998, x% lived in nursing homes by 2010"). my source for that 411? page 13 of the design doc. wicked. this new github repository contains five scripts: 1992 - 2010 download HRS microdata.R loop through every year and every file, download, then unzip everything in one big party impor t longitudinal RAND contributed files.R create a SQLite database (.db) on the local disk load the rand, rand-cams, and both rand-family files into the database (.db) in chunks (to prevent overloading ram) longitudinal RAND - analysis examples.R connect to the sql database created by the 'import longitudinal RAND contributed files' program create tw o database-backed complex sample survey object, using a taylor-series linearization design perform a mountain of analysis examples with wave weights from two different points in the panel import example HRS file.R load a fixed-width file using only the sas importation script directly into ram with < a href="http://blog.revolutionanalytics.com/2012/07/importing-public-data-with-sas-instructions-into-r.html">SAScii parse through the IF block at the bottom of the sas importation script, blank out a number of variables save the file as an R data file (.rda) for fast loading later replicate 2002 regression.R connect to the sql database created by the 'import longitudinal RAND contributed files' program create a database-backed complex sample survey object, using a taylor-series linearization design exactly match the final regression shown in this document provided by analysts at RAND as an update of the regression on pdf page B76 of this document . click here to view these five scripts for more detail about the health and retirement study (hrs), visit: michigan's hrs homepage rand's hrs homepage the hrs wikipedia page a running list of publications using hrs notes: exemplary work making it this far. as a reward, here's the detailed codebook for the main rand hrs file. note that rand also creates 'flat files' for every survey wave, but really, most every analysis you c an think of is possible using just the four files imported with the rand importation script above. if you must work with the non-rand files, there's an example of how to import a single hrs (umich-created) file, but if you wish to import more than one, you'll have to write some for loops yourself. confidential to sas, spss, stata, and sudaan users: a tidal wave is coming. you can get water up your nose and be dragged out to sea, or you can grab a surf board. time to transition to r. :D
Facebook
Twitterhttps://www.marketreportanalytics.com/privacy-policyhttps://www.marketreportanalytics.com/privacy-policy
The SAS (Serial Attached SCSI) interface tape library market is poised for substantial growth, projected to reach an estimated market size of $1,500 million by 2025. This expansion is driven by a confluence of factors, including the escalating volume of data generated across various industries and the inherent cost-effectiveness and reliability of tape storage for long-term archival and backup. Industries such as data centers, financial institutions, and government agencies are increasingly adopting SAS tape libraries to manage vast datasets efficiently, mitigate the risks associated with data loss, and comply with stringent data retention regulations. The market's Compound Annual Growth Rate (CAGR) is estimated to be robust, around 12%, underscoring the sustained demand for these solutions. Key applications like data archiving, disaster recovery, and compliance are expected to be the primary beneficiaries of this growth. Furthermore, the evolution of tape technology, offering increased storage density and faster transfer rates, continues to make SAS tape libraries a compelling choice compared to other storage media. The market landscape is characterized by a diverse range of players, from established giants like IBM and Quantum to emerging innovators, all vying for market share by offering a variety of tape library types, from small to large. This competitive environment fosters continuous innovation in terms of performance, capacity, and management software. Despite the rise of cloud storage, SAS tape libraries maintain a strong foothold, particularly for enterprise-level data management requiring offline, air-gapped, and tamper-proof storage solutions. Restraints such as the perceived complexity of implementation and the upfront investment for large-scale deployments are being addressed through improved software integration and service offerings. The market's trajectory is further influenced by emerging trends like the increasing adoption of Software-Defined Storage (SDS) and the growing emphasis on data security and ransomware protection, where tape libraries offer a unique advantage as an immutable storage medium. Here is a unique report description for SAS (Serial Attached SCSI) Interface Tape Libraries, incorporating your specifications:
Facebook
TwitterCompanies and individuals are storing increasingly more data digitally; however, much of the data is unused because it is unclassified. How many times have you opened your downloads folder, found a file you downloaded a year ago and you have no idea what the contents are? You can read through those files individually but imagine doing that for thousands of files. All that raw data in storage facilities create data lakes. As the amount of data grows and the complexity rises, data lakes become data swamps. The potentially valuable and interesting datasets will likely remain unused. Our tool addresses the need to classify these large pools of data in a visually effective and succinct manner by identifying keywords in datasets, and classifying datasets into a consistent taxonomy.
The files listed within kaggleDatasetSummaryTopicsClassification.csv have been processed with our tool to generate the keywords and taxonomic classification as seen below. The summaries are not generated from our system. Instead they were retrieved from user input as they uploaded the files on Kaggle. We planned to utilize these summaries to create an NLG model to generate summaries from any input file. Unfortunately we were not able to collect enough data to build a good model. Hopefully the data within this set might help future users achieve that goal.
Developed with Senior Design Center at NC State in collaboration with SAS. Senior Design Team: Tanya Chu, Katherine Marsh, Nikhil Milind, Anna Owens SAS Representatives: : Nancy Rausch, Marty Warner, Brant Kay, Tyler Wendell, JP Trawinski
Facebook
TwitterApache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
SAS-Bench: A Fine-Grained Benchmark for Evaluating Short Answer Scoring with Large Language Models
Dataset | ä¸ć–‡ | Paper | Code
🔍 Overview
SAS-Bench represents the first specialized benchmark for evaluating Large Language Models (LLMs) on Short Answer Scoring (SAS) tasks. Utilizing authentic questions from China's National College Entrance Examination (Gaokao)… See the full description on the dataset page: https://huggingface.co/datasets/aleversn/SAS-Bench.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
IntroductionA required step for presenting results of clinical studies is the declaration of participants demographic and baseline characteristics as claimed by the FDAAA 801. The common workflow to accomplish this task is to export the clinical data from the used electronic data capture system and import it into statistical software like SAS software or IBM SPSS. This software requires trained users, who have to implement the analysis individually for each item. These expenditures may become an obstacle for small studies. Objective of this work is to design, implement and evaluate an open source application, called ODM Data Analysis, for the semi-automatic analysis of clinical study data.MethodsThe system requires clinical data in the CDISC Operational Data Model format. After uploading the file, its syntax and data type conformity of the collected data is validated. The completeness of the study data is determined and basic statistics, including illustrative charts for each item, are generated. Datasets from four clinical studies have been used to evaluate the application’s performance and functionality.ResultsThe system is implemented as an open source web application (available at https://odmanalysis.uni-muenster.de) and also provided as Docker image which enables an easy distribution and installation on local systems. Study data is only stored in the application as long as the calculations are performed which is compliant with data protection endeavors. Analysis times are below half an hour, even for larger studies with over 6000 subjects.DiscussionMedical experts have ensured the usefulness of this application to grant an overview of their collected study data for monitoring purposes and to generate descriptive statistics without further user interaction. The semi-automatic analysis has its limitations and cannot replace the complex analysis of statisticians, but it can be used as a starting point for their examination and reporting.
Facebook
Twitterhttps://www.archivemarketresearch.com/privacy-policyhttps://www.archivemarketresearch.com/privacy-policy
The Predictive and Prescriptive Analytics Software market is experiencing robust growth, driven by the increasing adoption of data-driven decision-making across various industries. While precise figures for market size and CAGR are unavailable, considering the presence of major players like Microsoft, IBM, Oracle, and SAP, and the pervasive trend towards AI and machine learning, a reasonable estimation would place the 2025 market size at approximately $15 billion. Given the rapid advancements in analytics technologies and expanding data volumes, a conservative Compound Annual Growth Rate (CAGR) of 15% is projected from 2025 to 2033. This growth is fueled by several key factors: the burgeoning need for real-time insights in fast-paced business environments, the increasing availability of large datasets, and the declining cost and improving accessibility of advanced analytics tools. Organizations are increasingly leveraging predictive and prescriptive analytics to optimize operations, improve customer experiences, enhance risk management, and gain a competitive edge. This market is segmented across several key areas, including industry verticals (finance, healthcare, retail, manufacturing, etc.), deployment models (cloud, on-premise), and analytics types (predictive modeling, optimization, simulation). The competitive landscape is fiercely contested, with established tech giants and specialized analytics vendors vying for market share. While challenges such as data security concerns, integration complexities, and the need for skilled professionals exist, the overall market trajectory indicates continued expansion and further innovation in the years to come. The market's future is promising, with continuous development of more sophisticated algorithms and wider adoption across industries, promising further significant growth in the coming decade.
Facebook
Twitter
Facebook
TwitterIntroductionThis study aimed to investigate the possible associations between problematic smartphone use and brain functions in terms of both static and dynamic functional connectivity patterns.Materials and methodsResting-state functional magnetic resonance imaging data were scanned from 53 young healthy adults, all of whom completed the Short Version of the Smartphone Addiction Scale (SAS-SV) to assess their problematic smartphone use severity. Both static and dynamic functional brain network measures were evaluated for each participant. The brain network measures were correlated the SAS-SV scores, and compared between participants with and without a problematic smartphone use after adjusting for sex, age, education, and head motion.ResultsTwo participants were excluded because of excessive head motion, and 56.9% (29/51) of the final analyzed participants were found to have a problematic smartphone use (SAS-SV scores ≥ 31 for males and ≥ 33 for females, as proposed in prior research). At the global network level, the SAS-SV score was found to be significantly positively correlated with the global efficiency and local efficiency of static brain networks, and negatively correlated with the temporal variability using the dynamic brain network model. Large-scale subnetwork analyses indicated that a higher SAS-SV score was significantly associated with higher strengths of static functional connectivity within the frontoparietal and cinguloopercular subnetworks, as well as a lower temporal variability of dynamic functional connectivity patterns within the attention subnetwork. However, no significant differences were found when directly comparing between the groups of participants with and without a problematic smartphone use.ConclusionOur results suggested that problematic smartphone use is associated with differences in both the static and dynamic brain network organizations in young adults. These findings may help to identify at-risk population for smartphone addiction and guide targeted interventions for further research. Nevertheless, it might be necessary to confirm our findings in a larger sample, and to investigate if a more applicable SAS-SV cutoff point is required for defining problematic smartphone use in young Chinese adults nowadays.
Facebook
TwitterThis database is a collection of maps created from the 28 SAS-2 observation files. The original observation files can be accessed within BROWSE by changing to the SAS2RAW database. For each of the SAS-2 observation files, the analysis package FADMAP was run and the resulting maps, plus GIF images created from these maps, were collected into this database. Each map is a 60 x 60 pixel FITS format image with 1 degree pixels. The user may reconstruct any of these maps within the captive account by running FADMAP from the command line after extracting a file from within the SAS2RAW database. The parameters used for selecting data for these product map files are embedded keywords in the FITS maps themselves. These parameters are set in FADMAP, and for the maps in this database are set as 'wide open' as possible. That is, except for selecting on each of 3 energy ranges, all other FADMAP parameters were set using broad criteria. To find more information about how to run FADMAP on the raw event's file, the user can access help files within the SAS2RAW database or can use the 'fhelp' facility from the command line to gain information about FADMAP. This is a service provided by NASA HEASARC .
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
In recent years, researchers spent an increasing amount of effort investigating technical debt, with quantitative methods, and in particular static analysis, being the most common approach to investigate such a topic.
However, quantitative studies are susceptible, to varying degrees, to external validity threats, which hinder the generalisation of their findings.
In response to this concern, researchers strive to expand the scope of their studies by incorporating a larger number of projects into their analyses. This practice is typically executed on a case-by-case basis, necessitating substantial data collection efforts that have to be repeated for each new study.
To address this issue, this paper presents an approach for tackling this problem and enabling researchers to study architectural smells, a well-known indicator of architectural technical debt, at a large scale. Specifically, we introduce a novel approach to a data collection pipeline that leverages Apache Airflow to continuously generate up-to-date, large-scale datasets with any static analysis tool.
Finally, we use the data collected through the pipeline to study the correlation between architectural smells and logical coupling in order to understand how smells influence maintenance efforts.
Facebook
TwitterThe Delta Neighborhood Physical Activity Study was an observational study designed to assess characteristics of neighborhood built environments associated with physical activity. It was an ancillary study to the Delta Healthy Sprouts Project and therefore included towns and neighborhoods in which Delta Healthy Sprouts participants resided. The 12 towns were located in the Lower Mississippi Delta region of Mississippi. Data were collected via electronic surveys between August 2016 and September 2017 using the Rural Active Living Assessment (RALA) tools and the Community Park Audit Tool (CPAT). Scale scores for the RALA Programs and Policies Assessment and the Town-Wide Assessment were computed using the scoring algorithms provided for these tools via SAS software programming. The Street Segment Assessment and CPAT do not have associated scoring algorithms and therefore no scores are provided for them. Because the towns were not randomly selected and the sample size is small, the data may not be generalizable to all rural towns in the Lower Mississippi Delta region of Mississippi. Dataset one contains data collected with the RALA Programs and Policies Assessment (PPA) tool. Dataset two contains data collected with the RALA Town-Wide Assessment (TWA) tool. Dataset three contains data collected with the RALA Street Segment Assessment (SSA) tool. Dataset four contains data collected with the Community Park Audit Tool (CPAT). [Note : title changed 9/4/2020 to reflect study name] Resources in this dataset:Resource Title: Dataset One RALA PPA Data Dictionary. File Name: RALA PPA Data Dictionary.csvResource Description: Data dictionary for dataset one collected using the RALA PPA tool.Resource Software Recommended: Microsoft Excel,url: https://products.office.com/en-us/excel Resource Title: Dataset Two RALA TWA Data Dictionary. File Name: RALA TWA Data Dictionary.csvResource Description: Data dictionary for dataset two collected using the RALA TWA tool.Resource Software Recommended: Microsoft Excel,url: https://products.office.com/en-us/excel Resource Title: Dataset Three RALA SSA Data Dictionary. File Name: RALA SSA Data Dictionary.csvResource Description: Data dictionary for dataset three collected using the RALA SSA tool.Resource Software Recommended: Microsoft Excel,url: https://products.office.com/en-us/excel Resource Title: Dataset Four CPAT Data Dictionary. File Name: CPAT Data Dictionary.csvResource Description: Data dictionary for dataset four collected using the CPAT.Resource Software Recommended: Microsoft Excel,url: https://products.office.com/en-us/excel Resource Title: Dataset One RALA PPA. File Name: RALA PPA Data.csvResource Description: Data collected using the RALA PPA tool.Resource Software Recommended: Microsoft Excel,url: https://products.office.com/en-us/excel Resource Title: Dataset Two RALA TWA. File Name: RALA TWA Data.csvResource Description: Data collected using the RALA TWA tool.Resource Software Recommended: Microsoft Excel,url: https://products.office.com/en-us/excel Resource Title: Dataset Three RALA SSA. File Name: RALA SSA Data.csvResource Description: Data collected using the RALA SSA tool.Resource Software Recommended: Microsoft Excel,url: https://products.office.com/en-us/excel Resource Title: Dataset Four CPAT. File Name: CPAT Data.csvResource Description: Data collected using the CPAT.Resource Software Recommended: Microsoft Excel,url: https://products.office.com/en-us/excel Resource Title: Data Dictionary. File Name: DataDictionary_RALA_PPA_SSA_TWA_CPAT.csvResource Description: This is a combined data dictionary from each of the 4 dataset files in this set.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Accumulated-Depreciation Time Series for Obiz Concept SAS. Obiz S.A. provides relationship marketing and customer loyalty solutions in France and internationally. It develops web platforms and custom-made apps; and provides data driven marketing, animation, and support and consulting solutions, as well as operates Obiz, a digital platform. The company also offers e-gift cards, and leisure and tickets through its online store. It serves large companies, associations, and federations. Obiz S.A. was incorporated in 2010 and is headquartered in Lyon, France.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Ordinary least squares and stepwise selection are widespread in behavioral science research; however, these methods are well known to encounter overfitting problems such that R2 and regression coefficients may be inflated while standard errors and p values may be deflated, ultimately reducing both the parsimony of the model and the generalizability of conclusions. More optimal methods for selecting predictors and estimating regression coefficients such as regularization methods (e.g., Lasso) have existed for decades, are widely implemented in other disciplines, and are available in mainstream software, yet, these methods are essentially invisible in the behavioral science literature while the use of sub optimal methods continues to proliferate. This paper discusses potential issues with standard statistical models, provides an introduction to regularization with specific details on both Lasso and its related predecessor ridge regression, provides an example analysis and code for running a Lasso analysis in R and SAS, and discusses limitations and related methods.
Facebook
TwitterCC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
analyze the current population survey (cps) annual social and economic supplement (asec) with r the annual march cps-asec has been supplying the statistics for the census bureau's report on income, poverty, and health insurance coverage since 1948. wow. the us census bureau and the bureau of labor statistics ( bls) tag-team on this one. until the american community survey (acs) hit the scene in the early aughts (2000s), the current population survey had the largest sample size of all the annual general demographic data sets outside of the decennial census - about two hundred thousand respondents. this provides enough sample to conduct state- and a few large metro area-level analyses. your sample size will vanish if you start investigating subgroups b y state - consider pooling multiple years. county-level is a no-no. despite the american community survey's larger size, the cps-asec contains many more variables related to employment, sources of income, and insurance - and can be trended back to harry truman's presidency. aside from questions specifically asked about an annual experience (like income), many of the questions in this march data set should be t reated as point-in-time statistics. cps-asec generalizes to the united states non-institutional, non-active duty military population. the national bureau of economic research (nber) provides sas, spss, and stata importation scripts to create a rectangular file (rectangular data means only person-level records; household- and family-level information gets attached to each person). to import these files into r, the parse.SAScii function uses nber's sas code to determine how to import the fixed-width file, then RSQLite to put everything into a schnazzy database. you can try reading through the nber march 2012 sas importation code yourself, but it's a bit of a proc freak show. this new github repository contains three scripts: 2005-2012 asec - download all microdata.R down load the fixed-width file containing household, family, and person records import by separating this file into three tables, then merge 'em together at the person-level download the fixed-width file containing the person-level replicate weights merge the rectangular person-level file with the replicate weights, then store it in a sql database create a new variable - one - in the data table 2012 asec - analysis examples.R connect to the sql database created by the 'download all microdata' progr am create the complex sample survey object, using the replicate weights perform a boatload of analysis examples replicate census estimates - 2011.R connect to the sql database created by the 'download all microdata' program create the complex sample survey object, using the replicate weights match the sas output shown in the png file below 2011 asec replicate weight sas output.png statistic and standard error generated from the replicate-weighted example sas script contained in this census-provided person replicate weights usage instructions document. click here to view these three scripts for more detail about the current population survey - annual social and economic supplement (cps-asec), visit: the census bureau's current population survey page the bureau of labor statistics' current population survey page the current population survey's wikipedia article notes: interviews are conducted in march about experiences during the previous year. the file labeled 2012 includes information (income, work experience, health insurance) pertaining to 2011. when you use the current populat ion survey to talk about america, subract a year from the data file name. as of the 2010 file (the interview focusing on america during 2009), the cps-asec contains exciting new medical out-of-pocket spending variables most useful for supplemental (medical spending-adjusted) poverty research. confidential to sas, spss, stata, sudaan users: why are you still rubbing two sticks together after we've invented the butane lighter? time to transition to r. :D