/// <summary> /// Initializes a new instance of the BigDataPoolResourceInfo class. /// </summary> /// <param name="location">The geo-location where the resource /// lives</param> /// <param name="id">Fully qualified resource Id for the resource. Ex - /// /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}</param> /// <param name="name">The name of the resource</param> /// <param name="type">The type of the resource. Ex- /// Microsoft.Compute/virtualMachines or /// Microsoft.Storage/storageAccounts.</param> /// <param name="tags">Resource tags.</param> /// <param name="provisioningState">The state of the Big Data /// pool.</param> /// <param name="autoScale">Auto-scaling properties</param> /// <param name="creationDate">The time when the Big Data pool was /// created.</param> /// <param name="autoPause">Auto-pausing properties</param> /// <param name="sparkEventsFolder">The Spark events folder</param> /// <param name="nodeCount">The number of nodes in the Big Data /// pool.</param> /// <param name="libraryRequirements">Library version /// requirements</param> /// <param name="sparkVersion">The Apache Spark version.</param> /// <param name="defaultSparkLogFolder">The default folder where Spark /// logs will be written.</param> /// <param name="nodeSize">The level of compute power that each node in /// the Big Data pool has. Possible values include: 'None', 'Small', /// 'Medium', 'Large'</param> /// <param name="nodeSizeFamily">The kind of nodes that the Big Data /// pool provides. Possible values include: 'None', /// 'MemoryOptimized'</param> public BigDataPoolResourceInfo(string location, string id = default(string), string name = default(string), string type = default(string), IDictionary <string, string> tags = default(IDictionary <string, string>), string provisioningState = default(string), AutoScaleProperties autoScale = default(AutoScaleProperties), System.DateTime?creationDate = default(System.DateTime?), AutoPauseProperties autoPause = default(AutoPauseProperties), string sparkEventsFolder = default(string), int?nodeCount = default(int?), LibraryRequirements libraryRequirements = default(LibraryRequirements), string sparkVersion = default(string), string defaultSparkLogFolder = default(string), string nodeSize = default(string), string nodeSizeFamily = default(string)) : base(location, id, name, type, tags) { ProvisioningState = provisioningState; AutoScale = autoScale; CreationDate = creationDate; AutoPause = autoPause; SparkEventsFolder = sparkEventsFolder; NodeCount = nodeCount; LibraryRequirements = libraryRequirements; SparkVersion = sparkVersion; DefaultSparkLogFolder = defaultSparkLogFolder; NodeSize = nodeSize; NodeSizeFamily = nodeSizeFamily; CustomInit(); }
/// <summary> /// Initializes a new instance of the BigDataPoolResourceInfo class. /// </summary> /// <param name="location">The geo-location where the resource /// lives</param> /// <param name="id">Fully qualified resource ID for the resource. Ex - /// /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}</param> /// <param name="name">The name of the resource</param> /// <param name="type">The type of the resource. E.g. /// "Microsoft.Compute/virtualMachines" or /// "Microsoft.Storage/storageAccounts"</param> /// <param name="tags">Resource tags.</param> /// <param name="provisioningState">The state of the Big Data /// pool.</param> /// <param name="autoScale">Auto-scaling properties</param> /// <param name="creationDate">The time when the Big Data pool was /// created.</param> /// <param name="autoPause">Auto-pausing properties</param> /// <param name="isComputeIsolationEnabled">Whether compute isolation /// is required or not.</param> /// <param name="haveLibraryRequirementsChanged">Whether library /// requirements changed.</param> /// <param name="sessionLevelPackagesEnabled">Whether session level /// packages enabled.</param> /// <param name="sparkEventsFolder">The Spark events folder</param> /// <param name="nodeCount">The number of nodes in the Big Data /// pool.</param> /// <param name="libraryRequirements">Library version /// requirements</param> /// <param name="sparkConfigProperties">Spark configuration file to /// specify additional properties</param> /// <param name="sparkVersion">The Apache Spark version.</param> /// <param name="defaultSparkLogFolder">The default folder where Spark /// logs will be written.</param> /// <param name="nodeSize">The level of compute power that each node in /// the Big Data pool has. Possible values include: 'None', 'Small', /// 'Medium', 'Large', 'XLarge', 'XXLarge', 'XXXLarge'</param> /// <param name="nodeSizeFamily">The kind of nodes that the Big Data /// pool provides. Possible values include: 'None', /// 'MemoryOptimized'</param> public BigDataPoolResourceInfo(string location, string id = default(string), string name = default(string), string type = default(string), IDictionary <string, string> tags = default(IDictionary <string, string>), string provisioningState = default(string), AutoScaleProperties autoScale = default(AutoScaleProperties), System.DateTime?creationDate = default(System.DateTime?), AutoPauseProperties autoPause = default(AutoPauseProperties), bool?isComputeIsolationEnabled = default(bool?), bool?haveLibraryRequirementsChanged = default(bool?), bool?sessionLevelPackagesEnabled = default(bool?), string sparkEventsFolder = default(string), int?nodeCount = default(int?), LibraryRequirements libraryRequirements = default(LibraryRequirements), LibraryRequirements sparkConfigProperties = default(LibraryRequirements), string sparkVersion = default(string), string defaultSparkLogFolder = default(string), string nodeSize = default(string), string nodeSizeFamily = default(string)) : base(location, id, name, type, tags) { ProvisioningState = provisioningState; AutoScale = autoScale; CreationDate = creationDate; AutoPause = autoPause; IsComputeIsolationEnabled = isComputeIsolationEnabled; HaveLibraryRequirementsChanged = haveLibraryRequirementsChanged; SessionLevelPackagesEnabled = sessionLevelPackagesEnabled; SparkEventsFolder = sparkEventsFolder; NodeCount = nodeCount; LibraryRequirements = libraryRequirements; SparkConfigProperties = sparkConfigProperties; SparkVersion = sparkVersion; DefaultSparkLogFolder = defaultSparkLogFolder; NodeSize = nodeSize; NodeSizeFamily = nodeSizeFamily; CustomInit(); }
/// <summary> /// Initializes a new instance of the BigDataPoolResourceInfo class. /// </summary> /// <param name="location">The geo-location where the resource /// lives</param> /// <param name="id">Fully qualified resource ID for the resource. Ex - /// /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}</param> /// <param name="name">The name of the resource</param> /// <param name="type">The type of the resource. E.g. /// "Microsoft.Compute/virtualMachines" or /// "Microsoft.Storage/storageAccounts"</param> /// <param name="tags">Resource tags.</param> /// <param name="provisioningState">The state of the Big Data /// pool.</param> /// <param name="autoScale">Auto-scaling properties</param> /// <param name="creationDate">The time when the Big Data pool was /// created.</param> /// <param name="autoPause">Auto-pausing properties</param> /// <param name="isComputeIsolationEnabled">Whether compute isolation /// is required or not.</param> /// <param name="sessionLevelPackagesEnabled">Whether session level /// packages enabled.</param> /// <param name="cacheSize">The cache size</param> /// <param name="dynamicExecutorAllocation">Dynamic Executor /// Allocation</param> /// <param name="sparkEventsFolder">The Spark events folder</param> /// <param name="nodeCount">The number of nodes in the Big Data /// pool.</param> /// <param name="libraryRequirements">Library version /// requirements</param> /// <param name="customLibraries">List of custom libraries/packages /// associated with the spark pool.</param> /// <param name="sparkConfigProperties">Spark configuration file to /// specify additional properties</param> /// <param name="sparkVersion">The Apache Spark version.</param> /// <param name="defaultSparkLogFolder">The default folder where Spark /// logs will be written.</param> /// <param name="nodeSize">The level of compute power that each node in /// the Big Data pool has. Possible values include: 'None', 'Small', /// 'Medium', 'Large', 'XLarge', 'XXLarge', 'XXXLarge'</param> /// <param name="nodeSizeFamily">The kind of nodes that the Big Data /// pool provides. Possible values include: 'None', 'MemoryOptimized', /// 'HardwareAcceleratedFPGA', 'HardwareAcceleratedGPU'</param> /// <param name="lastSucceededTimestamp">The time when the Big Data /// pool was updated successfully.</param> public BigDataPoolResourceInfo(string location, string id = default(string), string name = default(string), string type = default(string), IDictionary <string, string> tags = default(IDictionary <string, string>), string provisioningState = default(string), AutoScaleProperties autoScale = default(AutoScaleProperties), System.DateTime?creationDate = default(System.DateTime?), AutoPauseProperties autoPause = default(AutoPauseProperties), bool?isComputeIsolationEnabled = default(bool?), bool?sessionLevelPackagesEnabled = default(bool?), int?cacheSize = default(int?), DynamicExecutorAllocation dynamicExecutorAllocation = default(DynamicExecutorAllocation), string sparkEventsFolder = default(string), int?nodeCount = default(int?), LibraryRequirements libraryRequirements = default(LibraryRequirements), IList <LibraryInfo> customLibraries = default(IList <LibraryInfo>), SparkConfigProperties sparkConfigProperties = default(SparkConfigProperties), string sparkVersion = default(string), string defaultSparkLogFolder = default(string), string nodeSize = default(string), string nodeSizeFamily = default(string), System.DateTime?lastSucceededTimestamp = default(System.DateTime?)) : base(location, id, name, type, tags) { ProvisioningState = provisioningState; AutoScale = autoScale; CreationDate = creationDate; AutoPause = autoPause; IsComputeIsolationEnabled = isComputeIsolationEnabled; SessionLevelPackagesEnabled = sessionLevelPackagesEnabled; CacheSize = cacheSize; DynamicExecutorAllocation = dynamicExecutorAllocation; SparkEventsFolder = sparkEventsFolder; NodeCount = nodeCount; LibraryRequirements = libraryRequirements; CustomLibraries = customLibraries; SparkConfigProperties = sparkConfigProperties; SparkVersion = sparkVersion; DefaultSparkLogFolder = defaultSparkLogFolder; NodeSize = nodeSize; NodeSizeFamily = nodeSizeFamily; LastSucceededTimestamp = lastSucceededTimestamp; CustomInit(); }