You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Sometimes when random data is generated - e.g. 1000 rows, it is desired to have some duplicates in the data set. E.g. when random data for "workplace" is generated the user could desire to have duplicate values in the "company" field within the 1000 rows. Today the size of the sample set is unknown.
The generators should allow to set the size of the sample set of the generated data. If e.g. in 1000 rows the sample set is 100, then one would expect to have duplicates.
Preferred solution
Allow to define a limiting value for each field that is generated. E.g. using a list of dicts specifying the name/key of the field and the value of the limit.
Alternatives
Alternatively, but not so flexible, would be to set a value per generator, so that all fields generated all use the same limit.
The text was updated successfully, but these errors were encountered:
Problem description
Sometimes when random data is generated - e.g. 1000 rows, it is desired to have some duplicates in the data set. E.g. when random data for "workplace" is generated the user could desire to have duplicate values in the "company" field within the 1000 rows. Today the size of the sample set is unknown.
The generators should allow to set the size of the sample set of the generated data. If e.g. in 1000 rows the sample set is 100, then one would expect to have duplicates.
Preferred solution
Allow to define a limiting value for each field that is generated. E.g. using a list of dicts specifying the name/key of the field and the value of the limit.
Alternatives
Alternatively, but not so flexible, would be to set a value per generator, so that all fields generated all use the same limit.
The text was updated successfully, but these errors were encountered: