Dataset Viewer
Auto-converted to Parquet Duplicate
Id
stringlengths
1
6
PostTypeId
stringclasses
6 values
AcceptedAnswerId
stringlengths
2
6
ParentId
stringlengths
1
6
Score
stringlengths
1
3
ViewCount
stringlengths
1
6
Body
stringlengths
0
32.5k
Title
stringlengths
15
150
ContentLicense
stringclasses
2 values
FavoriteCount
stringclasses
2 values
CreationDate
stringlengths
23
23
LastActivityDate
stringlengths
23
23
LastEditDate
stringlengths
23
23
LastEditorUserId
stringlengths
1
6
OwnerUserId
stringlengths
1
6
Tags
sequence
5
1
null
null
9
909
I've always been interested in machine learning, but I can't figure out one thing about starting out with a simple "Hello World" example - how can I avoid hard-coding behavior? For example, if I wanted to "teach" a bot how to avoid randomly placed obstacles, I couldn't just use relative motion, because the obstacles mo...
How can I do simple machine learning without hard-coding behavior?
CC BY-SA 3.0
null
2014-05-13T23:58:30.457
2014-05-14T00:36:31.077
null
null
5
[ "machine-learning" ]
7
1
10
null
4
483
As a researcher and instructor, I'm looking for open-source books (or similar materials) that provide a relatively thorough overview of data science from an applied perspective. To be clear, I'm especially interested in a thorough overview that provides material suitable for a college-level course, not particular piece...
What open-source books (or other materials) provide a relatively thorough overview of data science?
CC BY-SA 3.0
null
2014-05-14T00:11:06.457
2014-05-16T13:45:00.237
2014-05-16T13:45:00.237
97
36
[ "education", "open-source" ]
9
2
null
5
5
null
Not sure if this fits the scope of this SE, but here's a stab at an answer anyway. With all AI approaches you have to decide what it is you're modelling and what kind of uncertainty there is. Once you pick a framework that allows modelling of your situation, you then see which elements are "fixed" and which are flexibl...
null
CC BY-SA 3.0
null
2014-05-14T00:36:31.077
2014-05-14T00:36:31.077
null
null
51
null
10
2
null
7
13
null
One book that's freely available is "The Elements of Statistical Learning" by Hastie, Tibshirani, and Friedman (published by Springer): [see Tibshirani's website](http://statweb.stanford.edu/~tibs/ElemStatLearn/). Another fantastic source, although it isn't a book, is Andrew Ng's Machine Learning course on Coursera. Th...
null
CC BY-SA 3.0
null
2014-05-14T00:53:43.273
2014-05-14T00:53:43.273
null
null
22
null
14
1
29
null
26
1909
I am sure data science as will be discussed in this forum has several synonyms or at least related fields where large data is analyzed. My particular question is in regards to Data Mining. I took a graduate class in Data Mining a few years back. What are the differences between Data Science and Data Mining and in par...
Is Data Science the Same as Data Mining?
CC BY-SA 3.0
null
2014-05-14T01:25:59.677
2020-08-16T13:01:33.543
2014-06-17T16:17:20.473
322
66
[ "data-mining", "definitions" ]
15
1
null
null
2
656
In which situations would one system be preferred over the other? What are the relative advantages and disadvantages of relational databases versus non-relational databases?
What are the advantages and disadvantages of SQL versus NoSQL in data science?
CC BY-SA 3.0
null
2014-05-14T01:41:23.110
2014-05-14T01:41:23.110
null
null
64
[ "databases" ]
16
1
46
null
17
432
I use [Libsvm](http://www.csie.ntu.edu.tw/~cjlin/libsvm/) to train data and predict classification on semantic analysis problem. But it has a performance issue on large-scale data, because semantic analysis concerns n-dimension problem. Last year, [Liblinear](http://www.csie.ntu.edu.tw/~cjlin/liblinear/) was release, a...
Use liblinear on big data for semantic analysis
CC BY-SA 3.0
null
2014-05-14T01:57:56.880
2014-05-17T16:24:14.523
2014-05-17T16:24:14.523
84
63
[ "machine-learning", "bigdata", "libsvm" ]
17
5
null
null
0
null
[LIBSVM](http://www.csie.ntu.edu.tw/~cjlin/libsvm/) is a library for support vector classification (SVM) and regression. It was created by Chih-Chung Chang and Chih-Jen Lin in 2001.
null
CC BY-SA 3.0
null
2014-05-14T02:49:14.580
2014-05-16T13:44:53.470
2014-05-16T13:44:53.470
63
63
null
18
4
null
null
0
null
null
CC BY-SA 3.0
null
2014-05-14T02:49:14.580
2014-05-14T02:49:14.580
2014-05-14T02:49:14.580
-1
-1
null
19
1
37
null
94
19674
Lots of people use the term big data in a rather commercial way, as a means of indicating that large datasets are involved in the computation, and therefore potential solutions must have good performance. Of course, big data always carry associated terms, like scalability and efficiency, but what exactly defines a prob...
How big is big data?
CC BY-SA 3.0
null
2014-05-14T03:56:20.963
2018-05-01T13:04:43.563
2015-06-11T20:15:28.720
10119
84
[ "bigdata", "scalability", "efficiency", "performance" ]
20
1
26
null
19
434
We created a social network application for eLearning purposes. It's an experimental project that we are researching on in our lab. It has been used in some case studies for a while and the data in our relational DBMS (SQL Server 2008) is getting big. It's a few gigabytes now and the tables are highly connected to each...
The data in our relational DBMS is getting big, is it the time to move to NoSQL?
CC BY-SA 4.0
null
2014-05-14T05:37:46.780
2022-07-14T08:30:28.583
2019-09-07T18:23:57.040
29169
96
[ "nosql", "relational-dbms" ]
21
2
null
19
34
null
As you rightly note, these days "big data" is something everyone wants to say they've got, which entails a certain looseness in how people define the term. Generally, though, I'd say you're certainly dealing with big data if the scale is such that it's no longer feasible to manage with more traditional technologies su...
null
CC BY-SA 3.0
null
2014-05-14T05:44:29.340
2014-05-14T05:44:29.340
null
null
14
null
22
1
24
null
200
292233
My data set contains a number of numeric attributes and one categorical. Say, `NumericAttr1, NumericAttr2, ..., NumericAttrN, CategoricalAttr`, where `CategoricalAttr` takes one of three possible values: `CategoricalAttrValue1`, `CategoricalAttrValue2` or `CategoricalAttrValue3`. I'm using default [k-means clustering a...
K-Means clustering for mixed numeric and categorical data
CC BY-SA 4.0
null
2014-05-14T05:58:21.927
2022-10-14T09:40:25.270
2020-08-07T14:12:08.577
98307
97
[ "data-mining", "clustering", "octave", "k-means", "categorical-data" ]
23
2
null
7
8
null
Data Science specialization from Johns Hopkins University at Coursera would be a great start. [https://www.coursera.org/specialization/jhudatascience/1](https://www.coursera.org/specialization/jhudatascience/1)
null
CC BY-SA 3.0
null
2014-05-14T06:06:13.603
2014-05-14T06:06:13.603
null
null
97
null
24
2
null
22
181
null
The standard k-means algorithm isn't directly applicable to categorical data, for various reasons. The sample space for categorical data is discrete, and doesn't have a natural origin. A Euclidean distance function on such a space isn't really meaningful. As someone put it, "The fact a snake possesses neither wheels...
null
CC BY-SA 3.0
null
2014-05-14T06:26:27.163
2016-11-29T20:06:51.543
2016-11-29T20:06:51.543
14
14
null
25
2
null
19
7
null
Big Data is defined by the volume of data, that's right, but not only. The particularity of big data is that you need to store a lots of various and sometimes unstructured stuffs all the times and from a tons of sensors, usually for years or decade. Furthermore you need something scalable, so that it doesn't take you h...
null
CC BY-SA 3.0
null
2014-05-14T07:26:04.390
2014-05-14T07:26:04.390
null
null
104
null
26
2
null
20
15
null
A few gigabytes is not very "big". It's more like the normal size of an enterprise DB. As long as you go over PK when joining tables it should work out really well, even in the future (as long as you don't get TB's of data a day). Most professionals working in a big data environment consider > ~5TB as the beginning of ...
null
CC BY-SA 3.0
null
2014-05-14T07:38:31.103
2014-05-14T11:03:51.577
2014-05-14T11:03:51.577
115
115
null
27
2
null
20
10
null
To answer this question you have to answer which kind of compromise you can afford. RDBMs implements [ACID](http://en.wikipedia.org/wiki/ACID). This is expensive in terms of resources. There are no NoSQL solutions which are ACID. See [CAP theorem](http://en.wikipedia.org/wiki/CAP_theorem) to dive deep into these ideas....
null
CC BY-SA 3.0
null
2014-05-14T07:53:02.560
2014-05-14T08:03:37.890
2014-05-14T08:03:37.890
14
108
null
28
2
null
7
6
null
There is free ebook "[Introduction to Data Science](http://jsresearch.net/)" based on [r](/questions/tagged/r) language
null
CC BY-SA 3.0
null
2014-05-14T07:55:40.133
2014-05-14T07:55:40.133
null
null
118
null
29
2
null
14
27
null
[@statsRus](https://datascience.stackexchange.com/users/36/statsrus) starts to lay the groundwork for your answer in another question [What characterises the difference between data science and statistics?](https://datascience.meta.stackexchange.com/q/86/98307): > Data collection: web scraping and online surveys Data...
null
CC BY-SA 4.0
null
2014-05-14T07:56:34.437
2020-08-16T13:01:33.543
2020-08-16T13:01:33.543
98307
53
null
30
2
null
19
22
null
Total amount of data in the world: 2.8 zetabytes in 2012, estimated to reach 8 zetabytes by 2015 ([source](http://siliconangle.com/blog/2012/05/21/when-will-the-world-reach-8-zetabytes-of-stored-data-infographic/)) and with a doubling time of 40 months. Can't get bigger than that :) As an example of a single large orga...
null
CC BY-SA 3.0
null
2014-05-14T08:03:28.117
2014-05-14T18:30:59.180
2014-05-14T18:30:59.180
26
26
null
31
1
72
null
10
1760
I have a bunch of customer profiles stored in a [elasticsearch](/questions/tagged/elasticsearch) cluster. These profiles are now used for creation of target groups for our email subscriptions. Target groups are now formed manually using elasticsearch faceted search capabilities (like get all male customers of age 23 ...
Clustering customer data stored in ElasticSearch
CC BY-SA 3.0
null
2014-05-14T08:38:07.007
2022-10-21T03:12:52.913
2014-05-15T05:49:39.140
24
118
[ "data-mining", "clustering" ]
33
2
null
20
6
null
Is it the time to move to NoSQL will depends on 2 things: - The nature/structure of your data - Your current performance SQL databases excel when the data is well structured (e.g. when it can be modeled as a table, an Excel spreadsheet, or a set of rows with a fixed number of columns). Also good when you need to d...
null
CC BY-SA 3.0
null
2014-05-14T09:34:15.477
2017-08-29T11:26:37.137
2017-08-29T11:26:37.137
132
132
null
35
1
null
null
21
732
In working on exploratory data analysis, and developing algorithms, I find that most of my time is spent in a cycle of visualize, write some code, run on small dataset, repeat. The data I have tends to be computer vision/sensor fusion type stuff, and algorithms are vision-heavy (for example object detection and track...
How to scale up algorithm development?
CC BY-SA 3.0
null
2014-05-14T09:51:54.753
2014-05-20T03:56:43.147
null
null
26
[ "algorithms" ]
37
2
null
19
93
null
To me (coming from a relational database background), "Big Data" is not primarily about the data size (which is the bulk of what the other answers are so far). "Big Data" and "Bad Data" are closely related. Relational Databases require 'pristine data'. If the data is in the database, it is accurate, clean, and 100% rel...
null
CC BY-SA 3.0
null
2014-05-14T10:41:23.823
2018-05-01T13:04:43.563
2018-05-01T13:04:43.563
51450
9
null
38
1
43
null
15
3538
I heard about many tools / frameworks for helping people to process their data (big data environment). One is called Hadoop and the other is the noSQL concept. What is the difference in point of processing? Are they complementary?
What is the difference between Hadoop and noSQL
CC BY-SA 3.0
null
2014-05-14T10:44:58.933
2015-05-18T12:30:19.497
2014-05-14T22:26:59.453
134
134
[ "nosql", "tools", "processing", "apache-hadoop" ]
40
2
null
20
8
null
Big Data is actually not so about the "how big it is". First, few gigabytes is not big at all, it's almost nothing. So don't bother yourself, your system will continu to work efficiently for some time I think. Then you have to think of how do you use your data. - SQL approach: Every data is precious, well collected ...
null
CC BY-SA 3.0
null
2014-05-14T11:12:03.880
2014-05-14T11:12:03.880
null
null
104
null
41
1
44
null
55
10254
R has many libraries which are aimed at Data Analysis (e.g. JAGS, BUGS, ARULES etc..), and is mentioned in popular textbooks such as: J.Krusche, Doing Bayesian Data Analysis; B.Lantz, "Machine Learning with R". I've seen a guideline of 5TB for a dataset to be considered as Big Data. My question is: Is R suitable for th...
Is the R language suitable for Big Data
CC BY-SA 3.0
null
2014-05-14T11:15:40.907
2019-02-23T11:34:41.513
2014-05-14T13:06:28.407
118
136
[ "bigdata", "r" ]
42
2
null
38
5
null
NoSQL is a way to store data that does not require there to be some sort of relation. The simplicity of its design and horizontal scale-ability, one way they store data is the `key : value` pair design. This lends itself to processing that is similar to Hadoop. The use of a NoSQL db really depends on the type of proble...
null
CC BY-SA 3.0
null
2014-05-14T11:21:31.500
2014-05-14T11:21:31.500
null
null
59
null
43
2
null
38
16
null
Hadoop is not a database, hadoop is an entire ecosystem. ![the hadoop ecosystem](https://i.stack.imgur.com/oOYp7.png) Most people will refer to [mapreduce](http://de.wikipedia.org/wiki/MapReduce) jobs while talking about hadoop. A mapreduce job splits big datasets in some little chunks of data and spread them over a cl...
null
CC BY-SA 3.0
null
2014-05-14T11:23:25.913
2015-05-18T12:30:19.497
2015-05-18T12:30:19.497
115
115
null
44
2
null
41
43
null
Actually this is coming around. In the book R in a Nutshell there is even a section on using R with Hadoop for big data processing. There are some work arounds that need to be done because R does all it's work in memory, so you are basically limited to the amount of RAM you have available to you. A mature project for R...
null
CC BY-SA 3.0
null
2014-05-14T11:24:39.530
2015-01-31T11:34:03.700
2015-01-31T11:34:03.700
2522
59
null
45
2
null
35
7
null
First off, if your data has as many variations (in function of time, context, and others) as to make it hard to apply a single strategy to cope with it, you may be interested in doing a prior temporal/contextual/... characterization of the dataset. Characterizing data, i.e., extracting information about how the volume ...
null
CC BY-SA 3.0
null
2014-05-14T11:26:40.580
2014-05-14T11:33:27.770
2014-05-14T11:33:27.770
84
84
null
46
2
null
16
11
null
Note that there is an early version of LIBLINEAR ported to [Apache Spark](http://spark.apache.org). See [mailing list comments](http://apache-spark-user-list.1001560.n3.nabble.com/Spark-LIBLINEAR-td5546.html) for some early details, and the [project site](http://www.csie.ntu.edu.tw/~cjlin/libsvmtools/distributed-liblin...
null
CC BY-SA 3.0
null
2014-05-14T12:32:29.503
2014-05-14T21:03:05.313
2014-05-14T21:03:05.313
21
21
null
47
2
null
41
31
null
The main problem with using R for large data sets is the RAM constraint. The reason behind keeping all the data in RAM is that it provides much faster access and data manipulations than would storing on HDDs. If you are willing to take a hit on performance, then yes, it is quite practical to work with [large datasets i...
null
CC BY-SA 3.0
null
2014-05-14T12:39:41.197
2014-05-14T12:39:41.197
null
null
62
null
48
5
null
null
0
null
[R](http://www.r-project.org) is a language and environment for statistical computing and graphics. It is a GNU project which is similar to the S language and environment which was developed at Bell Laboratories (formerly AT&T, now Lucent Technologies) by John Chambers and colleagues. R can be considered as a different...
null
CC BY-SA 3.0
null
2014-05-14T13:08:26.647
2014-08-16T17:29:43.517
2014-08-16T17:29:43.517
2961
201
null
49
4
null
null
0
null
R is a free, open-source programming language and software environment for statistical computing, bioinformatics, and graphics.
null
CC BY-SA 3.0
null
2014-05-14T13:08:26.647
2014-08-15T16:38:27.880
2014-08-15T16:38:27.880
2961
2961
null
50
1
53
null
5
702
I have an R script that generates a report based on the current contents of a database. This database is constantly in flux with records being added/deleted many times each day. How can I ask my computer to run this every night at 4 am so that I have an up to date report waiting for me in the morning? Or perhaps I want...
Running an R script programmatically
CC BY-SA 3.0
null
2014-05-14T14:26:54.313
2014-05-14T15:42:02.393
null
null
151
[ "r", "databases", "efficiency", "tools" ]
51
2
null
50
9
null
> How can I ask my computer to run this every night at 4 am so that I have an up to date report waiting for me in the morning? You can set up a cronjob on a Linux system. These are run at the set time, if the computer is on. To do so, open a terminal and type: ``` crontab -e ``` And add: ``` 00 4 * * * r source(/ho...
null
CC BY-SA 3.0
null
2014-05-14T14:48:32.180
2014-05-14T14:57:32.243
2017-05-23T12:38:53.587
-1
62
null
52
1
null
null
36
4755
From my limited dabbling with data science using R, I realized that cleaning bad data is a very important part of preparing data for analysis. Are there any best practices or processes for cleaning data before processing it? If so, are there any automated or semi-automated tools which implement some of these best prac...
Organized processes to clean data
CC BY-SA 3.0
null
2014-05-14T15:25:21.700
2017-01-23T06:27:10.313
2014-05-16T13:45:07.447
136
157
[ "r", "data-cleaning" ]
53
2
null
50
10
null
For windows, use the task scheduler to set the task to run for example daily at 4:00 AM It gives you many other options regarding frequency etc. [http://en.wikipedia.org/wiki/Windows_Task_Scheduler](http://en.wikipedia.org/wiki/Windows_Task_Scheduler)
null
CC BY-SA 3.0
null
2014-05-14T15:42:02.393
2014-05-14T15:42:02.393
null
null
116
null
57
2
null
52
18
null
From my point of view, this question is suitable for a two-step answer. The first part, let us call it soft preprocessing, could be taken as the usage of different data mining algorithms to preprocess data in such a way that makes it suitable for further analyses. Notice that this could be the analysis itself, in case ...
null
CC BY-SA 3.0
null
2014-05-14T16:29:39.927
2014-05-14T16:29:39.927
null
null
84
null
58
2
null
16
10
null
You can check out [vowpal wabbit](http://hunch.net/~vw/). It is quite popular for large-scale learning and includes parallel provisions. From their website: > VW is the essence of speed in machine learning, able to learn from terafeature datasets with ease. Via parallel learning, it can exceed the throughput of any si...
null
CC BY-SA 3.0
null
2014-05-14T17:06:33.337
2014-05-14T17:06:33.337
null
null
119
null
59
1
316
null
11
1313
In reviewing “[Applied Predictive Modeling](http://rads.stackoverflow.com/amzn/click/1461468485)" a [reviewer states](http://www.information-management.com/blogs/applied-predictive-modeling-10024771-1.html): > One critique I have of statistical learning (SL) pedagogy is the absence of computation performance conside...
What are R's memory constraints?
CC BY-SA 3.0
null
2014-05-14T17:48:21.240
2014-07-26T15:10:51.000
2014-07-26T15:10:51.000
62
158
[ "apache-hadoop", "r" ]
60
2
null
59
8
null
R performs all computation in-memory so you can't perform operation on a dataset that is larger than available RAM amount. However there are some libraries that allow bigdata processing using R and one of popular libraries for bigdata processing like Hadoop.
null
CC BY-SA 3.0
null
2014-05-14T17:58:48.297
2014-05-14T17:58:48.297
null
null
118
null
61
1
62
null
56
16700
Logic often states that by overfitting a model, its capacity to generalize is limited, though this might only mean that overfitting stops a model from improving after a certain complexity. Does overfitting cause models to become worse regardless of the complexity of data, and if so, why is this the case? --- Related...
Why Is Overfitting Bad in Machine Learning?
CC BY-SA 3.0
null
2014-05-14T18:09:01.940
2017-09-17T02:27:31.110
2017-04-13T12:50:41.230
-1
158
[ "machine-learning", "predictive-modeling" ]
62
2
null
61
49
null
Overfitting is empirically bad. Suppose you have a data set which you split in two, test and training. An overfitted model is one that performs much worse on the test dataset than on training dataset. It is often observed that models like that also in general perform worse on additional (new) test datasets than mode...
null
CC BY-SA 3.0
null
2014-05-14T18:27:56.043
2015-02-12T07:08:27.463
2015-02-12T07:08:27.463
26
26
null
64
2
null
61
18
null
Overfitting, in a nutshell, means take into account too much information from your data and/or prior knowledge, and use it in a model. To make it more straightforward, consider the following example: you're hired by some scientists to provide them with a model to predict the growth of some kind of plants. The scientist...
null
CC BY-SA 3.0
null
2014-05-14T18:37:52.333
2014-05-15T23:22:39.427
2014-05-15T23:22:39.427
84
84
null
65
5
null
null
0
null
null
CC BY-SA 3.0
null
2014-05-14T18:45:23.917
2014-05-14T18:45:23.917
2014-05-14T18:45:23.917
-1
-1
null
66
4
null
null
0
null
Big data is the term for a collection of data sets so large and complex that it becomes difficult to process using on-hand database management tools or traditional data processing applications. The challenges include capture, curation, storage, search, sharing, transfer, analysis and visualization.
null
CC BY-SA 3.0
null
2014-05-14T18:45:23.917
2014-05-16T13:45:57.450
2014-05-16T13:45:57.450
118
118
null
67
5
null
null
0
null
GNU Octave is a high-level interpreted scripting language, primarily intended for numerical computations. It provides capabilities for the numerical solution of linear and nonlinear problems, and for performing other numerical experiments. It also provides extensive graphics capabilities for data visualization and mani...
null
CC BY-SA 4.0
null
2014-05-14T18:48:42.263
2019-04-08T17:28:07.320
2019-04-08T17:28:07.320
201
201
null
68
4
null
null
0
null
GNU Octave is a free and open-source mathematical software package and scripting language. The scripting language is intended to be compatible with MATLAB, but the two packages are not interchangeable. Don’t use both the [matlab] and [octave] tags, unless the question is explicitly about the similarities or differences...
null
CC BY-SA 4.0
null
2014-05-14T18:48:42.263
2019-04-08T17:28:15.870
2019-04-08T17:28:15.870
201
201
null
69
1
null
null
3
91
First, think it's worth me stating what I mean by replication & reproducibility: - Replication of analysis A results in an exact copy of all inputs and processes that are supply and result in incidental outputs in analysis B. - Reproducibility of analysis A results in inputs, processes, and outputs that are semantica...
Is it possible to automate generating reproducibility documentation?
CC BY-SA 3.0
null
2014-05-14T20:03:15.233
2014-05-15T02:02:08.010
2017-04-13T12:50:41.230
-1
158
[ "processing" ]
70
2
null
69
2
null
To be reproducible without being just a replication, you would need to redo the experiment with new data, following the same technique as before. The work flow is not as important as the techniques used. Sample data in the same way, use the same type of models. It doesn't matter if you switch from one language to an...
null
CC BY-SA 3.0
null
2014-05-14T22:03:50.597
2014-05-14T22:03:50.597
null
null
178
null
71
1
84
null
14
766
What are the data conditions that we should watch out for, where p-values may not be the best way of deciding statistical significance? Are there specific problem types that fall into this category?
When are p-values deceptive?
CC BY-SA 3.0
null
2014-05-14T22:12:37.203
2014-05-15T08:25:47.933
null
null
179
[ "bigdata", "statistics" ]
72
2
null
31
6
null
One algorithm that can be used for this is the [k-means clustering algorithm](http://en.wikipedia.org/wiki/K-means_clustering). Basically: - Randomly choose k datapoints from your set, $m_1$, ..., $m_k$. - Until convergence: Assign your data points to k clusters, where cluster i is the set of points for which m_i i...
null
CC BY-SA 4.0
null
2014-05-14T22:40:40.363
2022-10-21T03:12:52.913
2022-10-21T03:12:52.913
141355
22
null
73
2
null
71
5
null
You shouldn't consider the p-value out of context. One rather basic point (as illustrated by [xkcd](http://xkcd.com/882/)) is that you need to consider how many tests you're actually doing. Obviously, you shouldn't be shocked to see p < 0.05 for one out of 20 tests, even if the null hypothesis is true every time. A ...
null
CC BY-SA 3.0
null
2014-05-14T22:43:23.587
2014-05-14T22:43:23.587
null
null
14
null
74
2
null
71
2
null
One thing you should be aware of is the sample size you are using. Very large samples, such as economists using census data, will lead to deflated p-values. This paper ["Too Big to Fail: Large Samples and the p-Value Problem"](http://galitshmueli.com/system/files/Print%20Version.pdf) covers some of the issues.
null
CC BY-SA 3.0
null
2014-05-14T22:58:11.583
2014-05-14T22:58:11.583
null
null
64
null
75
1
78
null
5
168
If small p-values are plentiful in big data, what is a comparable replacement for p-values in data with million of samples?
Is there a replacement for small p-values in big data?
CC BY-SA 3.0
null
2014-05-15T00:26:11.387
2019-05-07T04:16:29.673
2019-05-07T04:16:29.673
1330
158
[ "statistics", "bigdata" ]
76
1
139
null
6
182
(Note: Pulled this question from the [list of questions in Area51](http://area51.stackexchange.com/proposals/55053/data-science/57398#57398), but believe the question is self explanatory. That said, believe I get the general intent of the question, and as a result likely able to field any questions on the question that...
Which Big Data technology stack is most suitable for processing tweets, extracting/expanding URLs and pushing (only) new links into 3rd party system?
CC BY-SA 3.0
null
2014-05-15T00:39:33.433
2014-05-18T15:18:08.050
2014-05-18T15:18:08.050
118
158
[ "bigdata", "tools", "data-stream-mining" ]
77
1
87
null
10
1217
Background: Following is from the book [Graph Databases](http://rads.stackoverflow.com/amzn/click/1449356265), which covers a performance test mentioned in the book [Neo4j in Action](http://rads.stackoverflow.com/amzn/click/1617290769): > Relationships in a graph naturally form paths. Querying, or traversing, the gr...
Is this Neo4j comparison to RDBMS execution time correct?
CC BY-SA 3.0
null
2014-05-15T01:22:35.167
2015-05-10T21:18:01.617
2014-05-15T13:15:02.727
118
158
[ "databases", "nosql", "neo4j" ]
78
2
null
75
7
null
There is no replacement in the strict sense of the word. Instead you should look at other measures. The other measures you look at depend on what you type of problem you are solving. In general, if you have a small p-value, also consider the magnitude of the effect size. It may be highly statistically significant bu...
null
CC BY-SA 3.0
null
2014-05-15T01:46:28.467
2014-05-15T01:46:28.467
2017-04-13T12:50:41.230
-1
178
null
79
5
null
null
0
null
Conceptually speaking, data-mining can be thought of as one item (or set of skills and applications) in the toolkit of the data scientist. More specifically, data-mining is an activity that seeks patterns in large, complex data sets. It usually emphasizes algorithmic techniques, but may also involve any set of related ...
null
CC BY-SA 3.0
null
2014-05-15T03:19:40.360
2017-08-27T17:25:18.230
2017-08-27T17:25:18.230
3117
53
null
80
4
null
null
0
null
An activity that seeks patterns in large, complex data sets. It usually emphasizes algorithmic techniques, but may also involve any set of related skills, applications, or methodologies with that goal.
null
CC BY-SA 3.0
null
2014-05-15T03:19:40.360
2014-05-16T13:46:05.850
2014-05-16T13:46:05.850
53
53
null
81
1
82
null
16
1433
What is(are) the difference(s) between parallel and distributed computing? When it comes to scalability and efficiency, it is very common to see solutions dealing with computations in clusters of machines, and sometimes it is referred to as a parallel processing, or as distributed processing. In a certain way, the comp...
Parallel and distributed computing
CC BY-SA 3.0
null
2014-05-15T04:59:54.317
2023-04-11T10:41:24.483
2014-05-15T09:31:51.370
118
84
[ "definitions", "parallel", "distributed" ]
82
2
null
81
17
null
Simply set, 'parallel' means running concurrently on distinct resources (CPUs), while 'distributed' means running across distinct computers, involving issues related to networks. Parallel computing using for instance [OpenMP](http://en.wikipedia.org/wiki/OpenMP) is not distributed, while parallel computing with [Messag...
null
CC BY-SA 3.0
null
2014-05-15T05:19:34.757
2014-05-15T05:25:39.970
2014-05-15T05:25:39.970
172
172
null
83
2
null
20
8
null
I posted a pretty detailed answer on stackoverflow about when it is appropriate to use relational vs document (or NoSQL) database, here: [Motivations for using relational database / ORM or document database / ODM](https://stackoverflow.com/questions/13528216/motivations-for-using-relational-database-orm-or-document-d...
null
CC BY-SA 3.0
null
2014-05-15T07:47:44.710
2014-05-15T07:59:05.497
2017-05-23T12:38:53.587
-1
26
null
84
2
null
71
10
null
You are asking about [Data Dredging](http://en.wikipedia.org/wiki/Data_dredging), which is what happens when testing a very large number of hypotheses against a data set, or testing hypotheses against a data set that were suggested by the same data. In particular, check out [Multiple hypothesis hazard](http://en.wiki...
null
CC BY-SA 3.0
null
2014-05-15T08:19:40.577
2014-05-15T08:25:47.933
2014-05-15T08:25:47.933
26
26
null
85
2
null
75
4
null
See also [When are p-values deceptive?](https://datascience.stackexchange.com/questions/71/when-are-p-values-deceptive/84#84) When there are a lot of variables that can be tested for pair-wise correlation (for example), the replacement is to use any of the corrections for [False discovery rate](http://en.wikipedia.org/...
null
CC BY-SA 3.0
null
2014-05-15T08:44:47.327
2014-05-15T20:32:26.923
2017-04-13T12:50:41.230
-1
26
null
86
1
101
null
15
2829
Given website access data in the form `session_id, ip, user_agent`, and optionally timestamp, following the conditions below, how would you best cluster the sessions into unique visitors? `session_id`: is an id given to every new visitor. It does not expire, however if the user doesn't accept cookies/clears cookies/cha...
Clustering unique visitors by useragent, ip, session_id
CC BY-SA 3.0
null
2014-05-15T09:04:09.710
2014-05-15T21:41:22.703
2014-05-15T10:06:06.393
116
116
[ "clustering" ]
87
2
null
77
8
null
Looking at this document called [Anatomy of Facebook](https://www.facebook.com/notes/facebook-data-team/anatomy-of-facebook/10150388519243859) I note that the median is 100. Looking at the cumulative function plot I can bet that the average is higher, near 200. So 50 seems to not be the best number here. However I thin...
null
CC BY-SA 3.0
null
2014-05-15T09:30:36.460
2014-05-15T09:30:36.460
null
null
108
null
89
1
91
null
12
544
For example, when searching something in Google, results return nigh-instantly. I understand that Google sorts and indexes pages with algorithms etc., but I imagine it infeasible for the results of every single possible query to be indexed (and results are personalized, which renders this even more infeasible)? Moreove...
How does a query into a huge database return with negligible latency?
CC BY-SA 3.0
null
2014-05-15T11:22:27.293
2014-08-30T18:40:02.403
2014-05-16T02:46:56.510
189
189
[ "bigdata", "google", "search" ]
90
2
null
81
4
null
The terms "parallel computing" and "distributed computing" certainly have a large overlap, but can be differentiated further. Actually, you already did this in your question, by later asking about "parallel processing" and "distributed processing". One could consider "distributed computing" as the more general term th...
null
CC BY-SA 3.0
null
2014-05-15T11:46:38.170
2014-05-15T11:46:38.170
null
null
156
null
91
2
null
89
13
null
Well, I'm not sure if it is MapReduce that solves the problem, but it surely wouldn't be MapReduce alone to solve all these questions you raised. But here are important things to take into account, and that make it feasible to have such low latency on queries from all these TBs of data in different machines: - distrib...
null
CC BY-SA 3.0
null
2014-05-15T11:56:43.607
2014-05-16T04:33:52.310
2014-05-16T04:33:52.310
84
84
null
92
2
null
89
10
null
MapReduce has nothing to do with real-time anything. It is a batch-oriented processing framework suitable for some offline tasks, like ETL and index building. Google has moved off of MapReduce for most jobs now, and even the Hadoop ecosystem is doing the same. The answer to low latency is generally to keep precomputed ...
null
CC BY-SA 3.0
null
2014-05-15T13:18:38.693
2014-05-15T13:18:38.693
null
null
21
null
93
2
null
86
6
null
There's not much you can do with just this data, but what little you can do does not rely on machine learning. Yes, sessions from the same IP but different User-Agents are almost certainly distinct users. Sessions with the same IP and User-Agent are usually the same user, except in the case of proxies / wi-fi access p...
null
CC BY-SA 3.0
null
2014-05-15T13:30:04.270
2014-05-15T13:30:04.270
null
null
21
null
94
1
97
null
21
301
While building a rank, say for a search engine, or a recommendation system, is it valid to rely on click frequency to determine the relevance of an entry?
Does click frequency account for relevance?
CC BY-SA 3.0
null
2014-05-15T14:41:24.020
2015-11-23T15:36:28.760
null
null
84
[ "recommender-system", "information-retrieval" ]
95
2
null
94
5
null
Is it valid to use click frequency, then yes. Is it valid to use only the click frequency, then probably no. Search relevance is much more complicated than just one metric. [There are entire books on the subject](http://www.amazon.ca/s/ref=nb_sb_noss?url=search-alias=aps&field-keywords=search%20ranking). Extending this...
null
CC BY-SA 3.0
null
2014-05-15T15:06:24.600
2014-05-15T15:06:24.600
null
null
9
null
96
2
null
94
7
null
For my part I can say that I use click frequency on i.e. eCommerce products. When you combine it with the days of the year it can even bring you great suggestions. i.e.: We have historical data from 1 year over 2 products (Snowboots[], Sandalettes[]) ``` Snowboots[1024,1253,652,123,50,12,8,4,50,148,345,896] Sandalettes...
null
CC BY-SA 3.0
null
2014-05-15T15:10:30.243
2015-11-23T15:36:28.760
2015-11-23T15:36:28.760
115
115
null
97
2
null
94
15
null
[Depends on the user's intent](http://research.microsoft.com/pubs/169639/cikm-clickpatterns.pdf), for starters. [Users normally only view the first set of links](http://www.seoresearcher.com/distribution-of-clicks-on-googles-serps-and-eye-tracking-analysis.htm), which means that unless the link is viewable, it's not g...
null
CC BY-SA 3.0
null
2014-05-15T17:14:36.817
2014-05-15T23:08:04.300
2014-05-15T23:08:04.300
158
158
null
101
2
null
86
9
null
One possibility here (and this is really an extension of what Sean Owen posted) is to define a "stable user." For the given info you have you can imagine making a user_id that is a hash of ip and some user agent info (pseudo code): ``` uid = MD5Hash(ip + UA.device + UA.model) ``` Then you flag these ids with "stable" ...
null
CC BY-SA 3.0
null
2014-05-15T21:41:22.703
2014-05-15T21:41:22.703
null
null
92
null
102
1
111
null
6
588
What is the best noSQL backend to use for a mobile game? Users can make a lot of servers requests, it needs also to retrieve users' historical records (like app purchasing) and analytics of usage behavior.
What is the Best NoSQL backend for a mobile game
CC BY-SA 3.0
null
2014-05-16T05:09:33.557
2016-12-09T21:55:46.000
2014-05-18T19:41:19.157
229
199
[ "nosql", "performance" ]
103
1
null
null
24
9503
Assume that we have a set of elements E and a similarity (not distance) function sim(ei, ej) between two elements ei,ej ∈ E. How could we (efficiently) cluster the elements of E, using sim? k-means, for example, requires a given k, Canopy Clustering requires two threshold values. What if we don't want such predefined ...
Clustering based on similarity scores
CC BY-SA 3.0
null
2014-05-16T14:26:12.270
2021-06-28T09:13:21.753
null
null
113
[ "clustering", "algorithms", "similarity" ]
104
5
null
null
0
null
## Use the definitions tag when: You think we should create an official definition. An existing Tag Wiki needs a more precise definition to avoid confusion and we need to create consensus before an edit. (rough draft - needs filling out)
null
CC BY-SA 3.0
null
2014-05-16T15:35:51.420
2014-05-20T13:50:52.447
2014-05-20T13:50:52.447
53
53
null
105
4
null
null
0
null
a discussion (meta) tag used when there exists *disagreement* or *confusion* about the everyday meaning of a term or phrase.
null
CC BY-SA 3.0
null
2014-05-16T15:35:51.420
2014-05-20T13:53:05.697
2014-05-20T13:53:05.697
53
53
null
106
2
null
14
6
null
There are many overlaps between data mining and datascience. I would say that people with the role of datamining are concerned with data collection and the extraction of features from unfiltered, unorganised and mostly raw/wild datasets. Some very important data may be difficult to extract, not do to the implementation...
null
CC BY-SA 3.0
null
2014-05-16T16:25:58.250
2014-05-16T16:25:58.250
null
null
34
null
107
1
null
null
12
146
Consider a stream containing [tuples](http://en.m.wikipedia.org/wiki/Tuple) `(user, new_score)` representing users' scores in an online game. The stream could have 100-1,000 new elements per second. The game has 200K to 300K unique players. I would like to have some standing queries like: - Which players posted more...
Opensource tools for help in mining stream of leader board scores
CC BY-SA 3.0
null
2014-05-16T20:07:50.983
2014-05-19T07:33:50.080
2014-05-19T07:33:50.080
118
200
[ "tools", "data-stream-mining" ]
108
5
null
null
1
null
null
CC BY-SA 3.0
null
2014-05-16T20:24:38.980
2014-05-16T20:24:38.980
2014-05-16T20:24:38.980
-1
-1
null
109
4
null
null
0
null
An activity that seeks patterns in a continuous stream of data elements, usually involving summarizing the stream in some way.
null
CC BY-SA 3.0
null
2014-05-16T20:24:38.980
2014-05-20T13:52:00.620
2014-05-20T13:52:00.620
200
200
null
111
2
null
102
7
null
Some factors you might consider: Developer familiarity: go with whatever you or your developers are familiar with. Mongo, Couch, Riak, DynamoDB etc all have their strengths but all should do ok here, so rather than going for an unfamiliar solution that might be slightly better go for familiar and save a bunch of devel...
null
CC BY-SA 3.0
null
2014-05-17T03:07:59.707
2014-05-18T06:43:52.453
2014-05-18T06:43:52.453
26
26
null
112
2
null
107
8
null
This isn't a full solution, but you may want to look into [OrientDB](http://www.orientechnologies.com/) as part of your stack. Orient is a Graph-Document database server written entirely in Java. In graph databases, relationships are considered first class citizens and therefore traversing those relationships can be d...
null
CC BY-SA 3.0
null
2014-05-17T04:18:10.020
2014-05-17T04:18:10.020
null
null
70
null
113
1
122
null
13
258
When a relational database, like MySQL, has better performance than a no relational, like MongoDB? I saw a question on Quora other day, about why Quora still uses MySQL as their backend, and that their performance is still good.
When a relational database has better performance than a no relational
CC BY-SA 3.0
null
2014-05-17T04:53:03.913
2017-06-05T19:30:23.440
2017-06-05T19:30:23.440
31513
199
[ "bigdata", "performance", "databases", "nosql" ]
115
1
131
null
15
4194
If I have a very long list of paper names, how could I get abstract of these papers from internet or any database? The paper names are like "Assessment of Utility in Web Mining for the Domain of Public Health". Does any one know any API that can give me a solution? I tried to crawl google scholar, however, google block...
Is there any APIs for crawling abstract of paper?
CC BY-SA 3.0
null
2014-05-17T08:45:08.420
2021-01-25T09:43:02.103
null
null
212
[ "data-mining", "machine-learning" ]
116
1
121
null
28
3243
I have a database from my Facebook application and I am trying to use machine learning to estimate users' age based on what Facebook sites they like. There are three crucial characteristics of my database: - the age distribution in my training set (12k of users in sum) is skewed towards younger users (i.e. I have 1157...
Machine learning techniques for estimating users' age based on Facebook sites they like
CC BY-SA 3.0
null
2014-05-17T09:16:18.823
2021-02-09T04:31:08.427
2014-05-17T19:26:53.783
173
173
[ "machine-learning", "dimensionality-reduction", "python" ]
118
5
null
null
0
null
NoSQL (sometimes expanded to "not only [sql](/questions/tagged/sql)") is a broad class of database management systems that differ from the classic model of the relational database management system ([rdbms](/questions/tagged/rdbms)) in some significant ways. ### NoSQL systems: - Specifically designed for high load ...
null
CC BY-SA 3.0
null
2014-05-17T13:41:20.283
2017-08-27T17:25:05.257
2017-08-27T17:25:05.257
381
201
null
119
4
null
null
0
null
NoSQL (sometimes expanded to "not only sql") is a broad class of database management systems that differ from the classic model of the relational database management system (rdbms) in some significant ways.
null
CC BY-SA 4.0
null
2014-05-17T13:41:20.283
2019-04-08T17:28:13.327
2019-04-08T17:28:13.327
201
201
null
120
2
null
115
5
null
arXiv has an [API and bulk download](http://arxiv.org/help/bulk_data) but if you want something for paid journals it will be hard to come by without paying an indexer like pubmed or elsevier or the like.
null
CC BY-SA 3.0
null
2014-05-17T18:15:11.937
2014-05-17T18:15:11.937
null
null
92
null
121
2
null
116
16
null
One thing to start off with would be k-NN. The idea here is that you have a user/item matrix and for some of the users you have a reported age. The age for a person in the user item matrix might be well determined by something like the mean or median age of some nearest neighbors in the item space. So you have each u...
null
CC BY-SA 3.0
null
2014-05-17T18:53:30.123
2014-05-17T18:53:30.123
null
null
92
null
122
2
null
113
10
null
It depends on your data and what you're doing with it. For example, if the processing you have to do requires transactions to synchronize across nodes, it will likely be faster to use transactions implemented in an RDBMS rather than implementing it yourself on top of NoSQL databases which don't support it natively.
null
CC BY-SA 3.0
null
2014-05-17T20:56:15.577
2014-05-17T20:56:15.577
null
null
180
null
123
5
null
null
0
null
The most basic relationship to describe is a linear relationship between variables, x and y, such that they can be said to be highly-correlated when every increase in x results in a proportional increase in y. They can also be said to be inversely proportional so that when x increases, y decreases. And finally, the two...
null
CC BY-SA 3.0
null
2014-05-17T21:10:41.990
2014-05-20T13:50:21.763
2014-05-20T13:50:21.763
53
53
null
124
4
null
null
0
null
A statistics term used to describe a type of dependence between variables (or data sets). Correlations are often used as an indicator of predictability. However, correlation does NOT imply causation. Different methods of calculating correlation exist to capture more complicated relationships between the variables being...
null
CC BY-SA 3.0
null
2014-05-17T21:10:41.990
2014-05-20T13:50:19.543
2014-05-20T13:50:19.543
53
53
null
End of preview. Expand in Data Studio

Dataset Card for "datascience-stackexchange-posts"

More Information needed

Downloads last month
9