diff --git a/content/en/blog/Volcano-1.11.0-release.md b/content/en/blog/Volcano-1.11.0-release.md
new file mode 100644
index 00000000..31134cb1
--- /dev/null
+++ b/content/en/blog/Volcano-1.11.0-release.md
@@ -0,0 +1,393 @@
++++
+title = "Volcano v1.11.0 Available Now"
+description = "New Features: Network topology-aware scheduling, Elastic hierarchical queues, Multi-cluster AI job scheduling, online/offline workloads colocation with dynamic resource overSubscription, Load-aware rescheduling, Fine-grained job fault recovery policies, Volcano Dashboard for resource visualization, Supports for Kubernetes v1.31, Volcano Job supports for Preemption Policy, and Performance optimizations for large-scale scenarios"
+subtitle = ""
+
+date = 2025-02-07
+lastmod = 2025-02-07
+datemonth = "Feb"
+dateyear = "2025"
+dateday = 07
+
+draft = false # Is this a draft? true/false
+toc = true # Show table of contents? true/false
+type = "posts" # Do not modify.
+authors = ["volcano"]
+
+tags = ["Practice"]
+summary = "New Features: Network topology-aware scheduling, Elastic hierarchical queues, Multi-cluster AI job scheduling, online/offline workloads colocation with dynamic resource overSubscription, Load-aware rescheduling, Fine-grained job fault recovery policies, Volcano Dashboard for resource visualization, Supports for Kubernetes v1.31, Volcano Job supports for Preemption Policy, and Performance optimizations for large-scale scenarios"
+
+# Add menu entry to sidebar.
+linktitle = "Volcano v1.11.0 Available Now"
+[menu.posts]
+parent = "tutorials"
+weight = 6
++++
+
+# Volcano v1.11 released: A New Era of Cloud-Native Scheduling for AI and Big Data
+
+As the de facto standard in cloud-native batch computing, Volcano has been widely adopted across various scenarios, including AI, Big Data, and High-Performance Computing (HPC). With over 800 contributors from more than 30 countries and tens of thousands of code commits, Volcano has been deployed in production environments by over 60 enterprises worldwide. It provides the industry with excellent practical standards and solutions for cloud native batch computing.
+
+As user scenarios grow increasingly complex, especially in the scenarios of vLLMs, there is a heightened demand for performance, GPU resource utilization, and availability in both training and inference workloads. This has driven Volcano to continuously expand its capabilities and address core user needs. Over the course of 28 releases, Volcano has introduced a series of enhancements and optimizations tailored to batch computing scenarios, helping users seamlessly migrate their workloads to cloud-native platforms. These improvements have resolved numerous pain points, earning the platform widespread praise and fostering a vibrant community with over 30 approvers and reviewers, creating a win-win ecosystem.
+
+The new release of Volcano will mark a new milestone in the New Year 2025, where the community will introduce a series of major features that will continue to deepen its focus on areas such as CNAI (Cloud Native AI) and Big Data, with key features including:
+
+**AI Scenarios:**
+
+- **Network Topology-Aware Scheduling:** Reduces network communication overhead between training tasks, optimizing performance for large AI model training.
+- **NPU Scheduling and Virtualization:** Enhances NPU resource utilization.
+- **GPU Dynamic Partitioning:** Introduces MIG and MPS dynamic partitioning to improve GPU resource utilization.
+- **Volcano Global Multi-Cluster AI Job Scheduling:** Supports Multi-cluster AI job deployment and distribution.
+- **Checkpointing and Fault Recovery Optimization:** Enables finer-grained job restart policies.
+- **Dynamic Resource Allocation (DRA):** Supports flexible and efficient management of heterogeneous resources.
+
+**Big Data Scenarios:**
+
+- **Elastic Hierarchical Queues:** Facilitates smooth migration of Big Data workloads to cloud-native platforms.
+
+**Microservices Scenarios:**
+
+- **Online and Offline Workloads colocation with Dynamic Resource OverSubscription:** Boosts resource utilization while ensuring QoS for online workloads.
+- **Load-Aware Scheduling and Descheduling:** Provides resource defragmentation and load balancing capabilities.
+
+**The official release of [Volcano v1.11](https://github.com/volcano-sh/volcano/releases/tag/v1.11.0)** marks a new chapter in cloud-native batch computing! This update focuses on the core needs of AI and Big Data, introducing **network topology-aware scheduling** and **multi-cluster AI job scheduling**, significantly enhancing the performance of AI training and inference tasks. Additionally, **online and offline workloads colocation with dynamic resource overcommitment** and **load-aware descheduling** further optimize resource utilization, ensuring high availability for online services. The introduction of **elastic hierarchical queues** offers more flexible scheduling strategies for Big Data scenarios.
+
+## **Deep Dive into Key Features**
+
+The v1.11 version released this time provides a series of major feature updates for AI, Big data and resource utilization improvement scenarios, mainly including:
+
+### **Network Topology-Aware Scheduling: Optimizing AI Large Model Training Performance**
+
+In AI large model training, model parallelism splits the model across multiple nodes, requiring frequent data exchange between nodes. Network communication often becomes a bottleneck, significantly impacting training efficiency. Data centers feature diverse network types like InfiniBand (IB), RoCE, and NVSwitch, with complex multi-layer switch topologies. The fewer switches spanned between two nodes, the lower the communication latency and the higher the throughput. Thus, users aim to schedule workloads in the optimal performance domain with the highest throughput and lowest latency.
+
+To address this, Volcano introduces **Network Topology-Aware Scheduling**, leveraging a unified network topology API and intelligent scheduling strategies to tackle network communication performance issues in large-scale AI training jobs.
+
+#### **Unified Network Topology API: Precise Network Structure Representation**
+
+To abstract away the differences in data center network types, Volcano defines a new CRD **HyperNode**, to represent network topology, providing a standardized API. Compared to traditional label-based approaches, HyperNode offers several advantages:
+
+- **Semantic Consistency:** HyperNode provides a standardized way to describe network topology, avoiding inconsistencies in label semantics.
+- **Hierarchical Structure:** HyperNode supports tree-like hierarchies, accurately reflecting actual network topologies.
+- **Ease of Management:** Cluster administrators can manually create HyperNodes or use automated network topology discovery tools to maintain them.
+
+A HyperNode represents a network topology performance domain, typically mapped to a switch. Multiple HyperNodes connect hierarchically to form a tree structure. For example:
+
+
{{}}
+
+
+- **Leaf HyperNodes** (s0, s1, s2, s3): Represent actual cluster nodes.
+- **Non-Leaf HyperNodes** (s4, s5, s6): Represent other HyperNodes.
+
+In this structure, communication efficiency between nodes depends on the number of HyperNode layers they span. For instance:
+
+- **node0** and **node1** within s0 have the highest communication efficiency.
+- **node1** and **node2** spanning two HyperNode layers (s0→s4→s1) have lower efficiency.
+- **node0** and **node4** spanning three HyperNode layers (s0→s4→s6) have the lowest efficiency.
+
+##### **HyperNode Configuration Example**
+
+Here’s an example of leaf and non-leaf HyperNode configurations:
+
+**Leaf HyperNode Example:**
+
+```yaml
+apiVersion: topology.volcano.sh/v1alpha1
+kind: HyperNode
+metadata:
+ name: s0
+spec:
+ tier: 1 # Lower tiers indicate higher communication efficiency
+ members: # List of child nodes
+ - type: Node # Child node type
+ selector:
+ exactMatch: # Exact match
+ name: node-0
+ - type: Node
+ selector:
+ regexMatch: # Regex match
+ pattern: node-[01]
+```
+
+**Non-Leaf HyperNode Example:**
+
+```yaml
+apiVersion: topology.volcano.sh/v1alpha1
+kind: HyperNode
+metadata:
+ name: s6
+spec:
+ tier: 3 # HyperNode tier
+ members: # List of child nodes
+ - type: HyperNode # Child node type
+ selector:
+ exactMatch: # Exact match
+ name: s4
+ - type: HyperNode
+ selector:
+ exactMatch: # Exact match
+ name: s5
+```
+
+#### **Network Topology-Aware Scheduling Strategy**
+
+Volcano Job and PodGroup can set topology constraints via the `networkTopology` field, supporting the following configurations:
+
+- **mode:** Supports `hard` and `soft` modes.
+ - `hard`: Enforces strict constraints, requiring tasks within a job to be deployed within the same HyperNode.
+ - `soft`: Prefers deploying tasks within the same HyperNode but allows flexibility.
+- **highestTierAllowed:** Used with `hard` mode to specify the maximum HyperNode tier a job can span.
+
+For example, the following configuration restricts a job to HyperNodes of tier 2 or lower (e.g., s4 or s5), otherwise, the job remains in a Pending state:
+
+```yaml
+spec:
+ networkTopology:
+ mode: hard
+ highestTierAllowed: 2
+```
+
+This scheduling strategy allows users to precisely control job topology constraints, ensuring optimal performance and significantly improving training efficiency.
+
+#### **Future Plans**
+
+Volcano will continue to refine network topology-aware scheduling, with future plans including:
+
+- Automating the conversion of node labels to HyperNode CRs to simplify migration.
+- Integrating network topology discovery tools to streamline HyperNode management.
+- Providing CLI tools for easier HyperNode hierarchy visualization and management.
+
+For detailed design and user guide, please refer to:
+
+Design Document: **[Network Topology Aware Scheduling](https://volcano.sh/en/docs/network_topology_aware_schedulin)**.
+
+Usage Document: **[Network Topology Aware Scheduling | Volcano](https://volcano.sh/en/docs/network_topology_aware_scheduling)**.
+
+Sincerely thanks to community developers: **@ecosysbin, @weapons97, @Xu-Wentao, @penggu, @JesseStutler, @Monokaix** for their contributions!
+
+### **Elastic Hierarchical Queues: Flexible Multi-Tenant Resource Management**
+
+In multi-tenant environments, fair resource allocation, isolation, and job prioritization are critical. Different departments or teams often share cluster resources while ensuring their jobs receive resources on demand, avoiding contention or waste. Volcano v1.11 introduces **Elastic Hierarchical Queues**, significantly enhancing queue resource management. Hierarchical queues enable finer-grained resource quota management, cross-level resource sharing and reclamation, and flexible preemption policies, creating an efficient and fair unified scheduling platform. For users migrating from YARN, Volcano seamlessly transitions Big Data workloads to Kubernetes clusters.
+
+#### **Core Capabilities of Elastic Hierarchical Queues**
+
+Volcano’s elastic hierarchical queues offer the following key features to meet complex multi-tenant demands:
+
+1. **Configurable Queue Hierarchies:** Users can create multi-level queues in a tree structure, each with independent resource quotas and priorities.
+2. **Cross-Level Resource Sharing and Reclamation:** Idle resources in child queues can be shared with sibling queues and reclaimed when needed.
+3. **Fine-Grained Resource Quota Management:** Each queue can set parameters like:
+ - `capability`: Maximum resource capacity.
+ - `deserved`: Fair share of resources; excess can be reclaimed.
+ - `guarantee`: Reserved resources, ensuring minimum guarantees.
+4. **Flexible Preemption Policies:** Supports priority-based preemption to ensure high-priority tasks receive resources promptly.
+
+#### **Hierarchical Queue Example**
+
+A simple hierarchical queue structure might look like this:
+
+{{}}
+
+- **Root Queue:** Manages global resource allocation.
+- **Department Queues:** Represent resource pools for different departments or teams.
+- **Child Queues:** Represent specific projects or tasks, where users submit jobs.
+
+#### **Use Cases**
+
+- **Multi-Department Resource Sharing:** Large enterprises can use hierarchical queues to fairly allocate and isolate resources across departments.
+- **Big Data Task Scheduling:** Users migrating from YARN to Kubernetes can leverage hierarchical queues for seamless Big Data workload migration.
+- **AI Training and Inference:** Hierarchical queues enable dynamic resource allocation and reclamation for AI tasks.
+
+For detailed design and user guide, please refer to:
+
+Design Document: **[hierarchical-queue-on-capacity-plugin](https://github.com/volcano-sh/volcano/blob/master/docs/design/hierarchical-queue-on-capacity-plugin.md)**.
+
+Usage Document: **[Hierarchical Queue | Volcano](https://volcano.sh/zh/docs/hierarchical_queue/)**.
+
+Sincerely thanks to community developer: **@Rui-Gan** for this contribution!
+
+### **Multi-Cluster AI Job Scheduling: Unified Management and Efficient Scheduling Across Clusters**
+
+As enterprise workloads grow, single Kubernetes clusters often fall short of meeting the demands of large-scale AI training and inference jobs. Users typically manage multiple Kubernetes clusters to achieve unified workload distribution, deployment, and management. Many users already deploy Volcano across multiple clusters, managed by **[Karmada](https://karmada.io/)**. To better support AI jobs in multi-cluster environments, Volcano has incubated the **[Volcano Global](https://github.com/volcano-sh/volcano-global)** sub-project, extending Volcano’s powerful scheduling capabilities to multi-cluster scenarios. This project provides a unified scheduling platform for multi-cluster AI jobs, supporting cross-cluster job distribution, resource management, and priority control.
+
+#### **Core Capabilities**
+
+Volcano Global enhances Karmada with the following features to meet the complex demands of multi-cluster AI job scheduling:
+
+1. **Cross-Cluster Volcano Job Scheduling:** Users can deploy and schedule Volcano Jobs across multiple clusters, maximizing resource utilization.
+2. **Queue Priority Scheduling:** Supports cross-cluster queue priority management, ensuring high-priority queues receive resources first.
+3. **Job Priority Scheduling and Queuing:** Enables job-level priority scheduling and queuing across clusters, ensuring critical tasks are executed promptly.
+4. **Multi-Tenant Fair Scheduling:** Provides fair resource allocation across tenants, preventing resource contention.
+
+
{{}}
+
+
+For detailed deployment and user guide, please refer to: **[Multi-Cluster AI Job Scheduling | Volcano](https://volcano.sh/en/docs/multi_cluster_scheduling/)**.
+
+Sincerely thanks to community developers: **@Vacant2333, @MondayCha, @lowang-bh, @Monokaix** for their contributions!
+
+### **Online and Offline Workloads colocation with Dynamic Resource OverSubscription: Maximizing Resource Utilization While Ensuring SLO**
+
+#### **Background: The Challenge of Resource Utilization**
+
+As cloud-native technologies advance, Kubernetes has become the "operating system" of the cloud-native era, with more workloads migrating to Kubernetes platforms. However, despite the flexibility and scalability of cloud-native technologies, data center resource utilization remains low. Online workloads (e.g., microservices) often exhibit peak-and-trough patterns, leaving resources idle during troughs and insufficient during peaks. To improve resource utilization while ensuring high-priority workload **SLOs (Service Level Objectives)**, Volcano introduces a **cloud-native colocation solution**, combining online and offline workloads with dynamic resource overSubscription to maximize cluster resource utilization while ensuring online workload stability.
+
+**Cloud-native colocation** involves deploying **online workloads** (e.g., real-time services) and **offline workloads** (e.g., batch jobs) on the same cluster. During online workload troughs, offline workloads utilize idle resources; during peaks, offline workloads are throttled to ensure online workload resource needs. This dynamic resource allocation mechanism not only improves resource utilization but also ensures online workload quality of service.
+
+#### **Industry Practices: Volcano’s Unique Advantages**
+
+While many companies have explored colocation technologies, existing solutions often fall short, such as being tightly coupled with Kubernetes, using rough overcommitment calculations, or offering inconsistent user experiences. Volcano addresses these issues with the following unique advantages:
+
+- **Native Support for Offline Job Scheduling:** Volcano Scheduler natively supports offline job scheduling without additional adaptation.
+- **Non-Invasive Design:** No modifications to Kubernetes are required, allowing users to adopt Volcano without altering existing cluster architectures.
+- **Dynamic Resource Overcommitment:** Real-time calculation of overcommittable resources ensures a balance between resource utilization and QoS.
+- **OS-Level Isolation and Guarantees:** Kernel-level resource isolation ensures online workload priority and stability.
+
+#### **Volcano Cloud-Native Colocation Solution: End-to-End Resource Optimization**
+
+Volcano’s cloud-native colocation solution provides end-to-end resource isolation and sharing mechanisms, including the following core components:
+
+**Volcano Scheduler:** Manages unified scheduling of online and offline workloads, offering abstractions like queues, groups, job priorities, fair scheduling, and resource reservations to meet the needs of microservices, Big Data, and AI workloads.
+
+**Volcano SLO Agent:** A daemonset running on each node, the SLO Agent monitors node resource usage, dynamically calculates resources that can be oversold, and allocates them to offline workloads. It also detects CPU/memory pressure and evicts offline workloads when necessary to ensure online workload priority.
+
+**Enhanced OS:** Volcano implements fine-grained QoS guarantees at the kernel level, using cgroups to set resource limits for online and offline workloads, ensuring online workloads receive sufficient resources even under high load.
+
+
{{}}
+Architecture
+
+
+#### **Core Capabilities: Balancing Resource Utilization and Stability**
+
+Volcano’s cloud-native colocation solution offers the following key capabilities to achieve both resource utilization and workload stability:
+
+- **Unified Scheduling:** Supports unified scheduling of microservices, batch and AI jobs.
+- **QoS-Based Resource Model:** Provides QoS-based resource management for online and offline workloads, ensuring high-priority workload stability.
+- **Dynamic Resource OverSubscription:** Dynamically calculates oversellable resources based on real-time CPU/memory usage, maximizing resource utilization.
+- **CPU Burst:** Allows containers to temporarily exceed CPU limits, avoiding throttling during critical moments and improving responsiveness.
+- **Network Bandwidth Isolation:** Supports node-level network bandwidth limits, ensuring online workload network requirements.
+
+For detailed design and user guide, please refer to: **[Cloud Native Colocation | Volcano](https://volcano.sh/en/docs/colocation/)**.
+
+Sincerely thanks to community developer: **@william-wang** for this contribution!
+
+### **Load-Aware Descheduling: Intelligent Cluster Resource Balancing**
+
+In Kubernetes clusters, dynamic workload changes often lead to uneven node resource utilization, causing hotspots that impact cluster stability and efficiency. To address this, Volcano v1.11 introduces **Load-Aware Rescheduling**, dynamically adjusting Pod distribution based on real node load to ensure balanced resource utilization and avoid hotspots, improving overall cluster performance and reliability. Load-aware rescheduling is incubated in the sub-project: https://github.com/volcano-sh/descheduler.
+
+
{{}}
+
+
+#### **Core Capabilities:**
+
+- **Load-Aware Scheduling:** Monitors real CPU and memory load metrics to dynamically adjust Pod distribution, avoiding reliance on Pod Request-based scheduling.
+- **Timed and Dynamic Triggers:** Supports CronTab-based or fixed-interval rescheduling to adapt to different scenarios.
+
+#### **Use Cases:**
+
+- **Uneven Node Resource Utilization:** Balances node load when some nodes are overutilized while others are underutilized.
+- **Hotspot Node Management:** Migrates Pods from overloaded nodes to ensure stability.
+
+For detailed user guide, please refer to: **[Load-aware Descheduling | Volcano](https://volcano.sh/en/docs/descheduler/)**.
+
+Sincerely thanks to community developer: **@Monokaix** for this contribution!
+
+### **Fine-Grained Job Fault Recovery: Efficient Task Interruption Handling**
+
+In AI, Big Data, and HPC scenarios, job stability and fault recovery are critical. Traditional fault recovery strategies often restart entire Jobs when a single Pod fails, wasting resources and potentially restarting training from scratch. With the rise of **checkpointing** and **resume-from-checkpoint** techniques, single Pod failures no longer require full Job restarts. Volcano v1.11 introduces **Fine-Grained Job Fault Recovery ** feature, offering flexible fault handling mechanisms to efficiently manage task interruptions and improve training efficiency.
+
+#### **Core Capabilities:**
+
+- **Pod-Level Restart:** Restarts only failed Pods, avoiding unnecessary Job restarts.
+- **Task-Level Restart:** Restarts the entire Task (a group of Pods) when a Pod fails, maintaining task group consistency.
+- **Timeout Mechanisms:** Sets timeouts for recovery actions to avoid overreacting to transient failures.
+
+For detailed design and user guide, please refer to: **[How to use job policy](https://github.com/volcano-sh/volcano/blob/master/docs/user-guide/how_to_use_job_policy.md)**.
+
+Sincerely thanks to community developer: **@bibibox** for this contribution!
+
+### **Volcano Dashboard: A Resource Visualization Component**
+
+The Volcano Dashboard is an official resource visualization component for Volcano. After deploying Volcano, users can deploy the dashboard to view and manage cluster resources through a graphical interface. The project is available at: https://github.com/volcano-sh/dashboard.
+
+Current features include:
+
+- Cluster overview, including Job counts, statuses, completion rates, Queue counts, and resource utilization.
+- Job and Queue lists with filtering, sorting, and search capabilities.
+- Pod lists with filtering, sorting, and search capabilities.
+
+Sincerely thanks to community developers: **@WY-Dev0, @Monokaix** for their contributions!
+
+### **Volcano Supports Kubernetes v1.31**
+
+Volcano closely follows Kubernetes releases, with full support for Kubernetes v1.31, including comprehensive UT and E2E testing to ensure functionality and reliability.
+
+To contribute to Volcano’s Kubernetes version adaptation, please refer to: **[adapt-k8s-todo](https://github.com/volcano-sh/volcano/blob/master/docs/design/adapt-k8s-todo.md)**.
+
+Sincerely thanks to community developers: **@vie-serendipity, @dongjiang1989** for their contributions!
+
+### **Volcano Job Supports Preemption Policy**
+
+Volcano Jobs now support **PreemptionPolicy**, allowing users to configure whether Jobs can preempt other Pods. Jobs with `PreemptionPolicy: Never` will not preempt resources, ensuring stability.
+
+For configuration examples, please refer to: **[how to configure priorityclass for job](https://github.com/volcano-sh/volcano/blob/master/docs/user-guide/how_to_configure_priorityclass_for_job.md)**.
+
+Sincerely thanks to community developer: **@JesseStut** for this contribution!
+
+### **Performance Optimization: Efficient Scheduling at Scale**
+
+In Volcano, **Queue** is one of the most fundamental and critical resources. The `status` field of a Queue records the states of PodGroups, such as `Unknown`, `Pending`, `Running`, `Inqueue`, and `Completed`. However, in large-scale scenarios, frequent changes in PodGroups within a Queue (e.g., when a large number of short-lived tasks are repeatedly submitted) can cause many PodGroups to transition from `Running` to `Completed`. In such cases, the Volcano Controller needs to frequently refresh the `status` field of the Queue, placing significant pressure on the APIServer. Additionally, the Volcano Scheduler updates the `status.allocated` field of the Queue after Job scheduling, which can lead to Queue update conflicts in large-scale environments, further impacting system performance.
+
+To completely resolve the issues of frequent Queue refreshes and update conflicts in large-scale scenarios, Volcano v1.11 has optimized the Queue management mechanism by migrating PodGroup statistics to **Metrics**, eliminating the need for persistent storage. This optimization significantly reduces the pressure on the APIServer while improving the overall performance and stability of the system.
+
+#### **Key Improvements After Optimization**
+
+**Migration of PodGroup Statistics to Metrics**
+PodGroup state data (e.g., `Unknown`, `Pending`, `Running`) is no longer stored in the `status` field of the Queue. Instead, it is recorded and displayed through the metrics system. Users can view the statistics of PodGroups in a Queue using the following commands:
+
+- **View statistics for a specific Queue**:
+
+ ```bash
+ vcctl queue get -n [name]
+ ```
+
+- **View statistics for all Queues**:
+
+ ```bash
+ vcctl queue list
+ ```
+
+**Reduced APIServer Pressure**
+By migrating PodGroup statistics to Metrics, frequent updates to the `status` field of the Queue are avoided, significantly reducing the load on the APIServer and improving system throughput.
+
+**Resolved Queue Update Conflicts**
+In large-scale scenarios, Queue update conflicts have been effectively mitigated, ensuring the efficient operation of the scheduler.
+
+For detailed design and metric names related to the migration of PodGroup state statistics to Metrics, please refer to: **[Queue podgroup statistics](https://github.com/volcano-sh/volcano/blob/master/docs/design/podgroup-statistics.md)**.
+
+Sincerely thanks to community developer: **@JesseStutler** for this contribution!
+
+## **Conclusion: Volcano v1.11, A New Era of Cloud-Native Batch Computing**
+
+Volcano v1.11 is not just a technological leap but a new chapter in cloud-native batch computing. Whether for AI large model training, Big Data scheduling, or resource optimization, Volcano v1.11 delivers powerful features and flexible solutions. We believe Volcano v1.11 will help users achieve greater heights in cloud-native batch computing, ushering in a new era of AI and Big Data scheduling!
+
+**Experience Volcano v1.11.0 now and step into a new era of efficient computing!**
+
+**v1.11.0 release:** https://github.com/volcano-sh/volcano/releases/tag/v1.11.0
+
+## **Acknowledgments**
+
+Volcano v1.11.0 includes contributions from 39 community members. Special thanks to all contributors:
+
+| @QingyaFan | @JesseStutler | @bogo-y |
+| :------------- | :-------------- | :--------------- |
+| @bibibox | @zedongh | @archlitchi |
+| @dongjiang1989 | @william-wang | @fengruotj |
+| @SataQiu | @lowang-bh | @Rui-Gan |
+| @xovoxy | @wangyang0616 | @PigNatovsky |
+| @Yanping-io | @lishangyuzi | @hwdef |
+| @bood | @kerthcet | @WY-Dev0 |
+| @raravena80 | @SherlockShemol | @zhifanggao |
+| @conghuhu | @MondayCha | @vie-serendipity |
+| @Prepmachine4 | @Monokaix | @lengrongfu |
+| @jasondrogba | @sceneryback | @TymonLee |
+| @liuyuanchun11 | @Vacant2333 | @matbme |
+| @lekaf974 | @kursataktas | @lut777 |
diff --git a/content/zh/blog/Volcano-1.11.0-release.md b/content/zh/blog/Volcano-1.11.0-release.md
new file mode 100644
index 00000000..f6024d63
--- /dev/null
+++ b/content/zh/blog/Volcano-1.11.0-release.md
@@ -0,0 +1,464 @@
++++
+title = "Volcano v1.11.0正式发布"
+description = "新增特性:网络拓扑感知调度、弹性层级队列、多集群AI作业调度、在离线混部与动态资源超卖、负载感知重调度、细粒度的作业故障恢复策略、Volcano Dashboard资源可视化、Volcano支持Kubernetes v1.31、Volcano Job支持Preemption Policy、大规模场景性能优化等"
+subtitle = ""
+
+date = 2025-02-07
+lastmod = 2025-02-07
+datemonth = "Feb"
+dateyear = "2025"
+dateday = 07
+
+draft = false # Is this a draft? true/false
+toc = true # Show table of contents? true/false
+type = "posts" # Do not modify.
+authors = ["volcano"]
+
+tags = ["Practice"]
+summary = "新增特性:网络拓扑感知调度、弹性层级队列、多集群AI作业调度、在离线混部与动态资源超卖、负载感知重调度、细粒度的作业故障恢复策略、Volcano Dashboard资源可视化、Volcano支持Kubernetes v1.31、Volcano Job支持Preemption Policy、大规模场景性能优化等"
+
+# Add menu entry to sidebar.
+linktitle = "Volcano v1.11.0正式发布"
+[menu.posts]
+parent = "tutorials"
+weight = 6
++++
+
+# Volcano v1.11 重磅发布!开启AI与大数据的云原生调度新纪元
+
+作为云原生批量计算领域的事实标准,Volcano已经在AI、Big Data及高性能计算 (HPC) 等多种场景中获得广泛应用,吸引了来自30多个国家的800多名贡献者,累计代码提交数万次。Volcano已在国内外60+企业进行了生产落地,经受住了实际生产环境的考验,赢得了用户的广泛赞誉,为业界提供了云原生批量计算的卓越实践标准与解决方案。
+
+随着用户使用场景的日益复杂,以及对资源利用率极致追求,特别是在AI大模型场景下,对训练与推理任务的性能、GPU资源利用率、可用性提出了更高的要求,促使Volcano不断拓展其应用场景,深入解决用户的核心诉求。Volcano目前的版本历程里共发布了28个release,针对批量计算的场景做了一系列功能增强和优化,帮助用户更好的将业务迁移到云原生平台,解决了诸多痛点问题,赢得了用户的广泛的喜爱与好评,用户与社区之间也形成了良好的互动,approver和reviewer数量累计发展了30+,达成了双赢互利的局面。
+
+值此2025新年之际,Volcano新版本将会是一个新的里程碑,社区将在2025年引入一系列重大特性,继续深耕CNAI(Cloud Native AI 云原生AI)和大数据等领域,主要特性包括:
+
+**AI场景:**
+
+- **网络拓扑感知调度:** 降低训练任务间的网络传输开销,优化大模型训练场景下的性能。
+- **NPU卡调度和虚拟化能力:** 提升NPU资源利用率。
+- **GPU卡动态切分能力:** 提供MIG与MPS动态切分能力,提升GPU资源利用率。
+- **Volcano Global多集群AI作业调度:** 支持跨集群的AI任务部署与拆分。
+- **断点续训与故障恢复能力优化:** 支持更细粒度的作业重启策略。
+- **支持DRA:**支持动态资源分配,灵活高效的管理异构资源。
+
+**大数据场景:**
+
+- **弹性层级队列能力:** 帮助用户将大数据业务丝滑迁移到云原生平台。
+
+**微服务场景:**
+
+- **在离线混部与动态资源超卖:** 提升资源利用率,同时保障在线业务QoS。
+- **负载感知调度与重调度:** 提供资源碎片整理和负载均衡能力。
+
+**[Volcano v1.11](https://github.com/volcano-sh/volcano/releases/tag/v1.11.0)的正式发布**,标志着云原生批量计算迈入全新阶段!本次更新聚焦AI与大数据的核心需求,推出**网络拓扑感知调度、多集群AI作业调度**等重磅特性,显著提升AI训练与推理任务的性能。同时,**在离线混部与动态资源超卖**及**负载感知重调度**功能进一步优化资源利用率,确保在线业务的高可用性。此外,**弹性层级队列**为大数据场景提供了更灵活的调度策略。Volcano v1.11不仅是技术的飞跃,更是云原生批量计算领域的全新标杆!
+
+## **重磅特性详解**
+
+本次发布的v1.11版本针对AI、大数据和资源利用率提升场景提供一系列重磅特性更新,主要包含:
+
+### **网络拓扑感知调度:优化AI大模型训练性能**
+
+在AI大模型训练场景中,模型并行(Model Parallelism)将模型分割到多个节点上,训练过程中这些节点需要频繁进行大量数据交互。此时,节点间的网络传输性能往往成为训练的瓶颈,显著影响训练效率。数据中心的网络类型多样,如InfiniBand (IB)、RoCE、NVSwitch等,且网络拓扑复杂,通常包含多层交换机。两个节点间跨的交换机越少,通信延迟越低,吞吐量越高。因此,用户希望将工作负载调度到具有最高吞吐量和最低延迟的最佳性能域,尽可能减少跨交换机的通信,以加速数据交换,提升训练效率。
+
+为此,Volcano提出了**网络拓扑感知调度(Network Topology Aware Scheduling)**策略,通过统一的网络拓扑API和智能调度策略,解决大规模数据中心AI训练任务的网络通信性能问题。
+
+#### **统一的网络拓扑API:精准表达网络结构**
+
+为了屏蔽数据中心网络类型的差异,Volcano定义了新的CRD **HyperNode**来表示网络拓扑,提供了标准化的API接口。与传统的通过节点标签(label)表示网络拓扑的方式相比,HyperNode具有以下优势:
+
+- **语义统一**:HyperNode提供了标准化的网络拓扑描述方式,避免了标签方式的语义不一致问题。
+- **层级结构**:HyperNode支持树状层级结构,能够更精确地表达实际的网络拓扑。
+- **易于管理**:集群管理员可以手动创建HyperNode,或通过网络拓扑自动发现工具维护HyperNode。
+
+一个HyperNode表示一个网络拓扑性能域,通常映射到一个交换机。多个HyperNode通过层级连接,形成树状结构。例如,下图展示了由多个HyperNode构成的网络拓扑:
+
+